go run hack/e2e.go -v --test --test_args='--ginkgo.focus=Kubernetes\se2e\ssuite\s\[It\]\s\[sig\-api\-machinery\]\sServers\swith\ssupport\sfor\sAPI\schunking\sshould\ssupport\scontinue\slisting\sfrom\sthe\slast\skey\sif\sthe\soriginal\sversion\shas\sbeen\scompacted\saway\,\sthough\sthe\slist\sis\sinconsistent\s\[Slow\]$'
test/e2e/framework/framework.go:241 k8s.io/kubernetes/test/e2e/framework.(*Framework).BeforeEach(0xc000980780) test/e2e/framework/framework.go:241 +0x96ffrom junit_01.xml
[BeforeEach] [sig-api-machinery] Servers with support for API chunking set up framework | framework.go:178 STEP: Creating a kubernetes client 11/26/22 05:33:00.329 Nov 26 05:33:00.329: INFO: >>> kubeConfig: /workspace/.kube/config STEP: Building a namespace api object, basename chunking 11/26/22 05:33:00.331 Nov 26 05:35:00.376: INFO: Unexpected error: <*fmt.wrapError | 0xc000dda000>: { msg: "wait for service account \"default\" in namespace \"chunking-6807\": timed out waiting for the condition", err: <*errors.errorString | 0xc0000d1da0>{ s: "timed out waiting for the condition", }, } Nov 26 05:35:00.376: FAIL: wait for service account "default" in namespace "chunking-6807": timed out waiting for the condition Full Stack Trace k8s.io/kubernetes/test/e2e/framework.(*Framework).BeforeEach(0xc000980780) test/e2e/framework/framework.go:241 +0x96f [AfterEach] [sig-api-machinery] Servers with support for API chunking test/e2e/framework/node/init/init.go:32 Nov 26 05:35:00.376: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready [DeferCleanup (Each)] [sig-api-machinery] Servers with support for API chunking dump namespaces | framework.go:196 STEP: dump namespace information after failure 11/26/22 05:35:00.423 STEP: Collecting events from namespace "chunking-6807". 11/26/22 05:35:00.423 STEP: Found 0 events. 11/26/22 05:35:00.463 Nov 26 05:35:00.505: INFO: POD NODE PHASE GRACE CONDITIONS Nov 26 05:35:00.505: INFO: Nov 26 05:35:00.547: INFO: Logging node info for node bootstrap-e2e-master Nov 26 05:35:00.589: INFO: Node Info: &Node{ObjectMeta:{bootstrap-e2e-master 9a48eb11-c964-4bf8-82c0-1c32a9703981 12024 0 2022-11-26 05:05:18 +0000 UTC <nil> <nil> map[beta.kubernetes.io/arch:amd64 beta.kubernetes.io/instance-type:n1-standard-1 beta.kubernetes.io/os:linux cloud.google.com/metadata-proxy-ready:true failure-domain.beta.kubernetes.io/region:us-west1 failure-domain.beta.kubernetes.io/zone:us-west1-b kubernetes.io/arch:amd64 kubernetes.io/hostname:bootstrap-e2e-master kubernetes.io/os:linux node.kubernetes.io/instance-type:n1-standard-1 topology.kubernetes.io/region:us-west1 topology.kubernetes.io/zone:us-west1-b] map[node.alpha.kubernetes.io/ttl:0 volumes.kubernetes.io/controller-managed-attach-detach:true] [] [] [{kubelet Update v1 2022-11-26 05:05:18 +0000 UTC FieldsV1 {"f:metadata":{"f:annotations":{".":{},"f:volumes.kubernetes.io/controller-managed-attach-detach":{}},"f:labels":{".":{},"f:beta.kubernetes.io/arch":{},"f:beta.kubernetes.io/instance-type":{},"f:beta.kubernetes.io/os":{},"f:cloud.google.com/metadata-proxy-ready":{},"f:failure-domain.beta.kubernetes.io/region":{},"f:failure-domain.beta.kubernetes.io/zone":{},"f:kubernetes.io/arch":{},"f:kubernetes.io/hostname":{},"f:kubernetes.io/os":{},"f:node.kubernetes.io/instance-type":{},"f:topology.kubernetes.io/region":{},"f:topology.kubernetes.io/zone":{}}},"f:spec":{"f:providerID":{},"f:unschedulable":{}}} } {kube-controller-manager Update v1 2022-11-26 05:05:25 +0000 UTC FieldsV1 {"f:status":{"f:conditions":{"k:{\"type\":\"NetworkUnavailable\"}":{"f:lastHeartbeatTime":{},"f:lastTransitionTime":{},"f:message":{},"f:reason":{},"f:status":{}}}}} status} {kube-controller-manager Update v1 2022-11-26 05:05:39 +0000 UTC FieldsV1 {"f:metadata":{"f:annotations":{"f:node.alpha.kubernetes.io/ttl":{}}},"f:spec":{"f:podCIDR":{},"f:podCIDRs":{".":{},"v:\"10.64.1.0/24\"":{}},"f:taints":{}}} } {kubelet Update v1 2022-11-26 05:32:43 +0000 UTC FieldsV1 {"f:status":{"f:conditions":{"k:{\"type\":\"DiskPressure\"}":{"f:lastHeartbeatTime":{}},"k:{\"type\":\"MemoryPressure\"}":{"f:lastHeartbeatTime":{}},"k:{\"type\":\"PIDPressure\"}":{"f:lastHeartbeatTime":{}},"k:{\"type\":\"Ready\"}":{"f:lastHeartbeatTime":{},"f:lastTransitionTime":{},"f:message":{},"f:reason":{},"f:status":{}}},"f:images":{}}} status}]},Spec:NodeSpec{PodCIDR:10.64.1.0/24,DoNotUseExternalID:,ProviderID:gce://k8s-jkns-e2e-gce-ci-reboot/us-west1-b/bootstrap-e2e-master,Unschedulable:true,Taints:[]Taint{Taint{Key:node-role.kubernetes.io/master,Value:,Effect:NoSchedule,TimeAdded:<nil>,},Taint{Key:node.kubernetes.io/unschedulable,Value:,Effect:NoSchedule,TimeAdded:<nil>,},},ConfigSource:nil,PodCIDRs:[10.64.1.0/24],},Status:NodeStatus{Capacity:ResourceList{attachable-volumes-gce-pd: {{127 0} {<nil>} 127 DecimalSI},cpu: {{1 0} {<nil>} 1 DecimalSI},ephemeral-storage: {{16656896000 0} {<nil>} 16266500Ki BinarySI},hugepages-1Gi: {{0 0} {<nil>} 0 DecimalSI},hugepages-2Mi: {{0 0} {<nil>} 0 DecimalSI},memory: {{3858374656 0} {<nil>} BinarySI},pods: {{110 0} {<nil>} 110 DecimalSI},},Allocatable:ResourceList{attachable-volumes-gce-pd: {{127 0} {<nil>} 127 DecimalSI},cpu: {{1 0} {<nil>} 1 DecimalSI},ephemeral-storage: {{14991206376 0} {<nil>} 14991206376 DecimalSI},hugepages-1Gi: {{0 0} {<nil>} 0 DecimalSI},hugepages-2Mi: {{0 0} {<nil>} 0 DecimalSI},memory: {{3596230656 0} {<nil>} BinarySI},pods: {{110 0} {<nil>} 110 DecimalSI},},Phase:,Conditions:[]NodeCondition{NodeCondition{Type:NetworkUnavailable,Status:False,LastHeartbeatTime:2022-11-26 05:05:25 +0000 UTC,LastTransitionTime:2022-11-26 05:05:25 +0000 UTC,Reason:RouteCreated,Message:RouteController created a route,},NodeCondition{Type:MemoryPressure,Status:False,LastHeartbeatTime:2022-11-26 05:32:43 +0000 UTC,LastTransitionTime:2022-11-26 05:05:18 +0000 UTC,Reason:KubeletHasSufficientMemory,Message:kubelet has sufficient memory available,},NodeCondition{Type:DiskPressure,Status:False,LastHeartbeatTime:2022-11-26 05:32:43 +0000 UTC,LastTransitionTime:2022-11-26 05:05:18 +0000 UTC,Reason:KubeletHasNoDiskPressure,Message:kubelet has no disk pressure,},NodeCondition{Type:PIDPressure,Status:False,LastHeartbeatTime:2022-11-26 05:32:43 +0000 UTC,LastTransitionTime:2022-11-26 05:05:18 +0000 UTC,Reason:KubeletHasSufficientPID,Message:kubelet has sufficient PID available,},NodeCondition{Type:Ready,Status:True,LastHeartbeatTime:2022-11-26 05:32:43 +0000 UTC,LastTransitionTime:2022-11-26 05:05:38 +0000 UTC,Reason:KubeletReady,Message:kubelet is posting ready status. AppArmor enabled,},},Addresses:[]NodeAddress{NodeAddress{Type:InternalIP,Address:10.138.0.2,},NodeAddress{Type:ExternalIP,Address:35.227.179.144,},NodeAddress{Type:InternalDNS,Address:bootstrap-e2e-master.c.k8s-jkns-e2e-gce-ci-reboot.internal,},NodeAddress{Type:Hostname,Address:bootstrap-e2e-master.c.k8s-jkns-e2e-gce-ci-reboot.internal,},},DaemonEndpoints:NodeDaemonEndpoints{KubeletEndpoint:DaemonEndpoint{Port:10250,},},NodeInfo:NodeSystemInfo{MachineID:0fdb3cfe29f66637553465718381a2f8,SystemUUID:0fdb3cfe-29f6-6637-5534-65718381a2f8,BootID:fb804a29-7e61-434d-8044-6944159ce35f,KernelVersion:5.10.123+,OSImage:Container-Optimized OS from Google,ContainerRuntimeVersion:containerd://1.7.0-beta.0-149-gd06318622,KubeletVersion:v1.27.0-alpha.0.50+70617042976dc1,KubeProxyVersion:v1.27.0-alpha.0.50+70617042976dc1,OperatingSystem:linux,Architecture:amd64,},Images:[]ContainerImage{ContainerImage{Names:[registry.k8s.io/kube-apiserver-amd64:v1.27.0-alpha.0.50_70617042976dc1],SizeBytes:135160272,},ContainerImage{Names:[registry.k8s.io/kube-controller-manager-amd64:v1.27.0-alpha.0.50_70617042976dc1],SizeBytes:124990265,},ContainerImage{Names:[registry.k8s.io/etcd@sha256:dd75ec974b0a2a6f6bb47001ba09207976e625db898d1b16735528c009cb171c registry.k8s.io/etcd:3.5.6-0],SizeBytes:102542580,},ContainerImage{Names:[registry.k8s.io/kube-scheduler-amd64:v1.27.0-alpha.0.50_70617042976dc1],SizeBytes:57660216,},ContainerImage{Names:[gke.gcr.io/prometheus-to-sd@sha256:e739643c3939ba0b161425f45a1989eedfc4a3b166db9a7100863296b4c70510 gke.gcr.io/prometheus-to-sd:v0.11.1-gke.1],SizeBytes:48742566,},ContainerImage{Names:[gcr.io/k8s-ingress-image-push/ingress-gce-glbc-amd64@sha256:5db27383add6d9f4ebdf0286409ac31f7f5d273690204b341a4e37998917693b gcr.io/k8s-ingress-image-push/ingress-gce-glbc-amd64:v1.20.1],SizeBytes:36598135,},ContainerImage{Names:[registry.k8s.io/addon-manager/kube-addon-manager@sha256:49cc4e6e4a3745b427ce14b0141476ab339bb65c6bc05033019e046c8727dcb0 registry.k8s.io/addon-manager/kube-addon-manager:v9.1.6],SizeBytes:30464183,},ContainerImage{Names:[registry.k8s.io/kas-network-proxy/proxy-server@sha256:2c111f004bec24888d8cfa2a812a38fb8341350abac67dcd0ac64e709dfe389c registry.k8s.io/kas-network-proxy/proxy-server:v0.0.33],SizeBytes:22020129,},ContainerImage{Names:[registry.k8s.io/metadata-proxy@sha256:e914645f22e946bce5165737e1b244e0a296ad1f0f81a9531adc57af2780978a registry.k8s.io/metadata-proxy:v0.1.12],SizeBytes:5301657,},ContainerImage{Names:[registry.k8s.io/pause@sha256:9001185023633d17a2f98ff69b6ff2615b8ea02a825adffa40422f51dfdcde9d registry.k8s.io/pause:3.8],SizeBytes:311286,},},VolumesInUse:[],VolumesAttached:[]AttachedVolume{},Config:nil,},} Nov 26 05:35:00.590: INFO: Logging kubelet events for node bootstrap-e2e-master Nov 26 05:35:00.634: INFO: Logging pods the kubelet thinks is on node bootstrap-e2e-master Nov 26 05:35:25.040: INFO: Unable to retrieve kubelet pods for node bootstrap-e2e-master: error trying to reach service: No agent available Nov 26 05:35:25.040: INFO: Logging node info for node bootstrap-e2e-minion-group-0975 Nov 26 05:35:25.082: INFO: Node Info: &Node{ObjectMeta:{bootstrap-e2e-minion-group-0975 6a0671d4-0e9d-4ac8-b359-65111262cfad 12294 0 2022-11-26 05:05:22 +0000 UTC <nil> <nil> map[beta.kubernetes.io/arch:amd64 beta.kubernetes.io/instance-type:n1-standard-2 beta.kubernetes.io/os:linux cloud.google.com/metadata-proxy-ready:true failure-domain.beta.kubernetes.io/region:us-west1 failure-domain.beta.kubernetes.io/zone:us-west1-b kubernetes.io/arch:amd64 kubernetes.io/hostname:bootstrap-e2e-minion-group-0975 kubernetes.io/os:linux node.kubernetes.io/instance-type:n1-standard-2 topology.hostpath.csi/node:bootstrap-e2e-minion-group-0975 topology.kubernetes.io/region:us-west1 topology.kubernetes.io/zone:us-west1-b] map[csi.volume.kubernetes.io/nodeid:{"csi-hostpath-multivolume-2344":"bootstrap-e2e-minion-group-0975","csi-hostpath-multivolume-5830":"bootstrap-e2e-minion-group-0975","csi-hostpath-multivolume-593":"bootstrap-e2e-minion-group-0975","csi-hostpath-provisioning-7492":"bootstrap-e2e-minion-group-0975"} node.alpha.kubernetes.io/ttl:0 volumes.kubernetes.io/controller-managed-attach-detach:true] [] [] [{kubelet Update v1 2022-11-26 05:05:22 +0000 UTC FieldsV1 {"f:metadata":{"f:annotations":{".":{},"f:volumes.kubernetes.io/controller-managed-attach-detach":{}},"f:labels":{".":{},"f:beta.kubernetes.io/arch":{},"f:beta.kubernetes.io/instance-type":{},"f:beta.kubernetes.io/os":{},"f:cloud.google.com/metadata-proxy-ready":{},"f:failure-domain.beta.kubernetes.io/region":{},"f:failure-domain.beta.kubernetes.io/zone":{},"f:kubernetes.io/arch":{},"f:kubernetes.io/hostname":{},"f:kubernetes.io/os":{},"f:node.kubernetes.io/instance-type":{},"f:topology.kubernetes.io/region":{},"f:topology.kubernetes.io/zone":{}}},"f:spec":{"f:providerID":{}}} } {kube-controller-manager Update v1 2022-11-26 05:05:23 +0000 UTC FieldsV1 {"f:metadata":{"f:annotations":{"f:node.alpha.kubernetes.io/ttl":{}}},"f:spec":{"f:podCIDR":{},"f:podCIDRs":{".":{},"v:\"10.64.2.0/24\"":{}}}} } {kube-controller-manager Update v1 2022-11-26 05:13:31 +0000 UTC FieldsV1 {"f:status":{"f:conditions":{"k:{\"type\":\"NetworkUnavailable\"}":{"f:lastHeartbeatTime":{},"f:lastTransitionTime":{},"f:message":{},"f:reason":{},"f:status":{}}}}} status} {node-problem-detector Update v1 2022-11-26 05:30:31 +0000 UTC FieldsV1 {"f:status":{"f:conditions":{"k:{\"type\":\"CorruptDockerOverlay2\"}":{".":{},"f:lastHeartbeatTime":{},"f:lastTransitionTime":{},"f:message":{},"f:reason":{},"f:status":{},"f:type":{}},"k:{\"type\":\"FrequentContainerdRestart\"}":{".":{},"f:lastHeartbeatTime":{},"f:lastTransitionTime":{},"f:message":{},"f:reason":{},"f:status":{},"f:type":{}},"k:{\"type\":\"FrequentDockerRestart\"}":{".":{},"f:lastHeartbeatTime":{},"f:lastTransitionTime":{},"f:message":{},"f:reason":{},"f:status":{},"f:type":{}},"k:{\"type\":\"FrequentKubeletRestart\"}":{".":{},"f:lastHeartbeatTime":{},"f:lastTransitionTime":{},"f:message":{},"f:reason":{},"f:status":{},"f:type":{}},"k:{\"type\":\"FrequentUnregisterNetDevice\"}":{".":{},"f:lastHeartbeatTime":{},"f:lastTransitionTime":{},"f:message":{},"f:reason":{},"f:status":{},"f:type":{}},"k:{\"type\":\"KernelDeadlock\"}":{".":{},"f:lastHeartbeatTime":{},"f:lastTransitionTime":{},"f:message":{},"f:reason":{},"f:status":{},"f:type":{}},"k:{\"type\":\"ReadonlyFilesystem\"}":{".":{},"f:lastHeartbeatTime":{},"f:lastTransitionTime":{},"f:message":{},"f:reason":{},"f:status":{},"f:type":{}}}}} status} {kubelet Update v1 2022-11-26 05:34:47 +0000 UTC FieldsV1 {"f:metadata":{"f:annotations":{"f:csi.volume.kubernetes.io/nodeid":{}},"f:labels":{"f:topology.hostpath.csi/node":{}}},"f:status":{"f:conditions":{"k:{\"type\":\"DiskPressure\"}":{"f:lastHeartbeatTime":{}},"k:{\"type\":\"MemoryPressure\"}":{"f:lastHeartbeatTime":{}},"k:{\"type\":\"PIDPressure\"}":{"f:lastHeartbeatTime":{}},"k:{\"type\":\"Ready\"}":{"f:lastHeartbeatTime":{},"f:lastTransitionTime":{},"f:message":{},"f:reason":{},"f:status":{}}},"f:images":{}}} status}]},Spec:NodeSpec{PodCIDR:10.64.2.0/24,DoNotUseExternalID:,ProviderID:gce://k8s-jkns-e2e-gce-ci-reboot/us-west1-b/bootstrap-e2e-minion-group-0975,Unschedulable:false,Taints:[]Taint{},ConfigSource:nil,PodCIDRs:[10.64.2.0/24],},Status:NodeStatus{Capacity:ResourceList{attachable-volumes-gce-pd: {{127 0} {<nil>} 127 DecimalSI},cpu: {{2 0} {<nil>} 2 DecimalSI},ephemeral-storage: {{101203873792 0} {<nil>} 98831908Ki BinarySI},hugepages-1Gi: {{0 0} {<nil>} 0 DecimalSI},hugepages-2Mi: {{0 0} {<nil>} 0 DecimalSI},memory: {{7815430144 0} {<nil>} BinarySI},pods: {{110 0} {<nil>} 110 DecimalSI},},Allocatable:ResourceList{attachable-volumes-gce-pd: {{127 0} {<nil>} 127 DecimalSI},cpu: {{2 0} {<nil>} 2 DecimalSI},ephemeral-storage: {{91083486262 0} {<nil>} 91083486262 DecimalSI},hugepages-1Gi: {{0 0} {<nil>} 0 DecimalSI},hugepages-2Mi: {{0 0} {<nil>} 0 DecimalSI},memory: {{7553286144 0} {<nil>} BinarySI},pods: {{110 0} {<nil>} 110 DecimalSI},},Phase:,Conditions:[]NodeCondition{NodeCondition{Type:CorruptDockerOverlay2,Status:False,LastHeartbeatTime:2022-11-26 05:30:31 +0000 UTC,LastTransitionTime:2022-11-26 05:05:27 +0000 UTC,Reason:NoCorruptDockerOverlay2,Message:docker overlay2 is functioning properly,},NodeCondition{Type:FrequentUnregisterNetDevice,Status:False,LastHeartbeatTime:2022-11-26 05:30:31 +0000 UTC,LastTransitionTime:2022-11-26 05:05:27 +0000 UTC,Reason:NoFrequentUnregisterNetDevice,Message:node is functioning properly,},NodeCondition{Type:FrequentKubeletRestart,Status:False,LastHeartbeatTime:2022-11-26 05:30:31 +0000 UTC,LastTransitionTime:2022-11-26 05:05:27 +0000 UTC,Reason:NoFrequentKubeletRestart,Message:kubelet is functioning properly,},NodeCondition{Type:FrequentDockerRestart,Status:False,LastHeartbeatTime:2022-11-26 05:30:31 +0000 UTC,LastTransitionTime:2022-11-26 05:05:27 +0000 UTC,Reason:NoFrequentDockerRestart,Message:docker is functioning properly,},NodeCondition{Type:FrequentContainerdRestart,Status:False,LastHeartbeatTime:2022-11-26 05:30:31 +0000 UTC,LastTransitionTime:2022-11-26 05:05:27 +0000 UTC,Reason:NoFrequentContainerdRestart,Message:containerd is functioning properly,},NodeCondition{Type:KernelDeadlock,Status:False,LastHeartbeatTime:2022-11-26 05:30:31 +0000 UTC,LastTransitionTime:2022-11-26 05:05:27 +0000 UTC,Reason:KernelHasNoDeadlock,Message:kernel has no deadlock,},NodeCondition{Type:ReadonlyFilesystem,Status:False,LastHeartbeatTime:2022-11-26 05:30:31 +0000 UTC,LastTransitionTime:2022-11-26 05:05:27 +0000 UTC,Reason:FilesystemIsNotReadOnly,Message:Filesystem is not read-only,},NodeCondition{Type:NetworkUnavailable,Status:False,LastHeartbeatTime:2022-11-26 05:05:39 +0000 UTC,LastTransitionTime:2022-11-26 05:05:39 +0000 UTC,Reason:RouteCreated,Message:RouteController created a route,},NodeCondition{Type:MemoryPressure,Status:False,LastHeartbeatTime:2022-11-26 05:34:35 +0000 UTC,LastTransitionTime:2022-11-26 05:05:22 +0000 UTC,Reason:KubeletHasSufficientMemory,Message:kubelet has sufficient memory available,},NodeCondition{Type:DiskPressure,Status:False,LastHeartbeatTime:2022-11-26 05:34:35 +0000 UTC,LastTransitionTime:2022-11-26 05:05:22 +0000 UTC,Reason:KubeletHasNoDiskPressure,Message:kubelet has no disk pressure,},NodeCondition{Type:PIDPressure,Status:False,LastHeartbeatTime:2022-11-26 05:34:35 +0000 UTC,LastTransitionTime:2022-11-26 05:05:22 +0000 UTC,Reason:KubeletHasSufficientPID,Message:kubelet has sufficient PID available,},NodeCondition{Type:Ready,Status:True,LastHeartbeatTime:2022-11-26 05:34:35 +0000 UTC,LastTransitionTime:2022-11-26 05:05:23 +0000 UTC,Reason:KubeletReady,Message:kubelet is posting ready status. AppArmor enabled,},},Addresses:[]NodeAddress{NodeAddress{Type:InternalIP,Address:10.138.0.5,},NodeAddress{Type:ExternalIP,Address:35.247.104.236,},NodeAddress{Type:InternalDNS,Address:bootstrap-e2e-minion-group-0975.c.k8s-jkns-e2e-gce-ci-reboot.internal,},NodeAddress{Type:Hostname,Address:bootstrap-e2e-minion-group-0975.c.k8s-jkns-e2e-gce-ci-reboot.internal,},},DaemonEndpoints:NodeDaemonEndpoints{KubeletEndpoint:DaemonEndpoint{Port:10250,},},NodeInfo:NodeSystemInfo{MachineID:b38385f7976a474936da82f5f565760a,SystemUUID:b38385f7-976a-4749-36da-82f5f565760a,BootID:0aa74273-9b8d-48ac-bad0-0a1805f010ea,KernelVersion:5.10.123+,OSImage:Container-Optimized OS from Google,ContainerRuntimeVersion:containerd://1.7.0-beta.0-149-gd06318622,KubeletVersion:v1.27.0-alpha.0.50+70617042976dc1,KubeProxyVersion:v1.27.0-alpha.0.50+70617042976dc1,OperatingSystem:linux,Architecture:amd64,},Images:[]ContainerImage{ContainerImage{Names:[registry.k8s.io/e2e-test-images/jessie-dnsutils@sha256:24aaf2626d6b27864c29de2097e8bbb840b3a414271bf7c8995e431e47d8408e registry.k8s.io/e2e-test-images/jessie-dnsutils:1.7],SizeBytes:112030336,},ContainerImage{Names:[registry.k8s.io/e2e-test-images/volume/nfs@sha256:3bda73f2428522b0e342af80a0b9679e8594c2126f2b3cca39ed787589741b9e registry.k8s.io/e2e-test-images/volume/nfs:1.3],SizeBytes:95836203,},ContainerImage{Names:[registry.k8s.io/sig-storage/nfs-provisioner@sha256:e943bb77c7df05ebdc8c7888b2db289b13bf9f012d6a3a5a74f14d4d5743d439 registry.k8s.io/sig-storage/nfs-provisioner:v3.0.1],SizeBytes:90632047,},ContainerImage{Names:[registry.k8s.io/kube-proxy-amd64:v1.27.0-alpha.0.50_70617042976dc1],SizeBytes:67201736,},ContainerImage{Names:[registry.k8s.io/e2e-test-images/agnhost@sha256:16bbf38c463a4223d8cfe4da12bc61010b082a79b4bb003e2d3ba3ece5dd5f9e registry.k8s.io/e2e-test-images/agnhost:2.43],SizeBytes:51706353,},ContainerImage{Names:[gke.gcr.io/prometheus-to-sd@sha256:e739643c3939ba0b161425f45a1989eedfc4a3b166db9a7100863296b4c70510 gke.gcr.io/prometheus-to-sd:v0.11.1-gke.1],SizeBytes:48742566,},ContainerImage{Names:[registry.k8s.io/e2e-test-images/httpd@sha256:148b022f5c5da426fc2f3c14b5c0867e58ef05961510c84749ac1fddcb0fef22 registry.k8s.io/e2e-test-images/httpd:2.4.38-4],SizeBytes:40764257,},ContainerImage{Names:[registry.k8s.io/metrics-server/metrics-server@sha256:6385aec64bb97040a5e692947107b81e178555c7a5b71caa90d733e4130efc10 registry.k8s.io/metrics-server/metrics-server:v0.5.2],SizeBytes:26023008,},ContainerImage{Names:[registry.k8s.io/sig-storage/csi-provisioner@sha256:ee3b525d5b89db99da3b8eb521d9cd90cb6e9ef0fbb651e98bb37be78d36b5b8 registry.k8s.io/sig-storage/csi-provisioner:v3.3.0],SizeBytes:25491225,},ContainerImage{Names:[registry.k8s.io/sig-storage/csi-resizer@sha256:425d8f1b769398127767b06ed97ce62578a3179bcb99809ce93a1649e025ffe7 registry.k8s.io/sig-storage/csi-resizer:v1.6.0],SizeBytes:24148884,},ContainerImage{Names:[registry.k8s.io/sig-storage/csi-snapshotter@sha256:291334908ddf71a4661fd7f6d9d97274de8a5378a2b6fdfeb2ce73414a34f82f registry.k8s.io/sig-storage/csi-snapshotter:v6.1.0],SizeBytes:23881995,},ContainerImage{Names:[registry.k8s.io/sig-storage/csi-attacher@sha256:9a685020911e2725ad019dbce6e4a5ab93d51e3d4557f115e64343345e05781b registry.k8s.io/sig-storage/csi-attacher:v4.0.0],SizeBytes:23847201,},ContainerImage{Names:[registry.k8s.io/sig-storage/hostpathplugin@sha256:92257881c1d6493cf18299a24af42330f891166560047902b8d431fb66b01af5 registry.k8s.io/sig-storage/hostpathplugin:v1.9.0],SizeBytes:18758628,},ContainerImage{Names:[registry.k8s.io/autoscaling/addon-resizer@sha256:43f129b81d28f0fdd54de6d8e7eacd5728030782e03db16087fc241ad747d3d6 registry.k8s.io/autoscaling/addon-resizer:1.8.14],SizeBytes:10153852,},ContainerImage{Names:[registry.k8s.io/sig-storage/csi-node-driver-registrar@sha256:0103eee7c35e3e0b5cd8cdca9850dc71c793cdeb6669d8be7a89440da2d06ae4 registry.k8s.io/sig-storage/csi-node-driver-registrar:v2.5.1],SizeBytes:9133109,},ContainerImage{Names:[registry.k8s.io/sig-storage/livenessprobe@sha256:933940f13b3ea0abc62e656c1aa5c5b47c04b15d71250413a6b821bd0c58b94e registry.k8s.io/sig-storage/livenessprobe:v2.7.0],SizeBytes:8688564,},ContainerImage{Names:[registry.k8s.io/kas-network-proxy/proxy-agent@sha256:48f2a4ec3e10553a81b8dd1c6fa5fe4bcc9617f78e71c1ca89c6921335e2d7da registry.k8s.io/kas-network-proxy/proxy-agent:v0.0.33],SizeBytes:8512162,},ContainerImage{Names:[registry.k8s.io/metadata-proxy@sha256:e914645f22e946bce5165737e1b244e0a296ad1f0f81a9531adc57af2780978a registry.k8s.io/metadata-proxy:v0.1.12],SizeBytes:5301657,},ContainerImage{Names:[registry.k8s.io/e2e-test-images/busybox@sha256:c318242786b139d18676b1c09a0ad7f15fc17f8f16a5b2e625cd0dc8c9703daf registry.k8s.io/e2e-test-images/busybox:1.29-2],SizeBytes:732424,},ContainerImage{Names:[registry.k8s.io/e2e-test-images/busybox@sha256:2e0f836850e09b8b7cc937681d6194537a09fbd5f6b9e08f4d646a85128e8937 registry.k8s.io/e2e-test-images/busybox:1.29-4],SizeBytes:731990,},ContainerImage{Names:[registry.k8s.io/pause@sha256:9001185023633d17a2f98ff69b6ff2615b8ea02a825adffa40422f51dfdcde9d registry.k8s.io/pause:3.8],SizeBytes:311286,},},VolumesInUse:[],VolumesAttached:[]AttachedVolume{},Config:nil,},} Nov 26 05:35:25.082: INFO: Logging kubelet events for node bootstrap-e2e-minion-group-0975 Nov 26 05:35:25.126: INFO: Logging pods the kubelet thinks is on node bootstrap-e2e-minion-group-0975 Nov 26 05:35:25.169: INFO: Unable to retrieve kubelet pods for node bootstrap-e2e-minion-group-0975: error trying to reach service: No agent available Nov 26 05:35:25.169: INFO: Logging node info for node bootstrap-e2e-minion-group-rwsn Nov 26 05:35:25.211: INFO: Node Info: &Node{ObjectMeta:{bootstrap-e2e-minion-group-rwsn adedf225-b462-4305-a60c-9042d31ca1d2 12343 0 2022-11-26 05:05:15 +0000 UTC <nil> <nil> map[beta.kubernetes.io/arch:amd64 beta.kubernetes.io/instance-type:n1-standard-2 beta.kubernetes.io/os:linux cloud.google.com/metadata-proxy-ready:true failure-domain.beta.kubernetes.io/region:us-west1 failure-domain.beta.kubernetes.io/zone:us-west1-b kubernetes.io/arch:amd64 kubernetes.io/hostname:bootstrap-e2e-minion-group-rwsn kubernetes.io/os:linux node.kubernetes.io/instance-type:n1-standard-2 topology.hostpath.csi/node:bootstrap-e2e-minion-group-rwsn topology.kubernetes.io/region:us-west1 topology.kubernetes.io/zone:us-west1-b] map[csi.volume.kubernetes.io/nodeid:{"csi-hostpath-multivolume-3900":"bootstrap-e2e-minion-group-rwsn","csi-hostpath-provisioning-8014":"bootstrap-e2e-minion-group-rwsn"} node.alpha.kubernetes.io/ttl:0 volumes.kubernetes.io/controller-managed-attach-detach:true] [] [] [{kubelet Update v1 2022-11-26 05:05:15 +0000 UTC FieldsV1 {"f:metadata":{"f:annotations":{".":{},"f:volumes.kubernetes.io/controller-managed-attach-detach":{}},"f:labels":{".":{},"f:beta.kubernetes.io/arch":{},"f:beta.kubernetes.io/instance-type":{},"f:beta.kubernetes.io/os":{},"f:cloud.google.com/metadata-proxy-ready":{},"f:failure-domain.beta.kubernetes.io/region":{},"f:failure-domain.beta.kubernetes.io/zone":{},"f:kubernetes.io/arch":{},"f:kubernetes.io/hostname":{},"f:kubernetes.io/os":{},"f:node.kubernetes.io/instance-type":{},"f:topology.kubernetes.io/region":{},"f:topology.kubernetes.io/zone":{}}},"f:spec":{"f:providerID":{}}} } {kube-controller-manager Update v1 2022-11-26 05:05:17 +0000 UTC FieldsV1 {"f:metadata":{"f:annotations":{"f:node.alpha.kubernetes.io/ttl":{}}},"f:spec":{"f:podCIDR":{},"f:podCIDRs":{".":{},"v:\"10.64.0.0/24\"":{}}}} } {kube-controller-manager Update v1 2022-11-26 05:29:28 +0000 UTC FieldsV1 {"f:status":{"f:conditions":{"k:{\"type\":\"NetworkUnavailable\"}":{"f:lastHeartbeatTime":{},"f:lastTransitionTime":{},"f:message":{},"f:reason":{},"f:status":{}}},"f:volumesAttached":{}}} status} {kubelet Update v1 2022-11-26 05:34:59 +0000 UTC FieldsV1 {"f:metadata":{"f:annotations":{"f:csi.volume.kubernetes.io/nodeid":{}},"f:labels":{"f:topology.hostpath.csi/node":{}}},"f:status":{"f:conditions":{"k:{\"type\":\"DiskPressure\"}":{"f:lastHeartbeatTime":{}},"k:{\"type\":\"MemoryPressure\"}":{"f:lastHeartbeatTime":{}},"k:{\"type\":\"PIDPressure\"}":{"f:lastHeartbeatTime":{}},"k:{\"type\":\"Ready\"}":{"f:lastHeartbeatTime":{},"f:lastTransitionTime":{},"f:message":{},"f:reason":{},"f:status":{}}},"f:images":{},"f:volumesInUse":{}}} status} {node-problem-detector Update v1 2022-11-26 05:35:22 +0000 UTC FieldsV1 {"f:status":{"f:conditions":{"k:{\"type\":\"CorruptDockerOverlay2\"}":{".":{},"f:lastHeartbeatTime":{},"f:lastTransitionTime":{},"f:message":{},"f:reason":{},"f:status":{},"f:type":{}},"k:{\"type\":\"FrequentContainerdRestart\"}":{".":{},"f:lastHeartbeatTime":{},"f:lastTransitionTime":{},"f:message":{},"f:reason":{},"f:status":{},"f:type":{}},"k:{\"type\":\"FrequentDockerRestart\"}":{".":{},"f:lastHeartbeatTime":{},"f:lastTransitionTime":{},"f:message":{},"f:reason":{},"f:status":{},"f:type":{}},"k:{\"type\":\"FrequentKubeletRestart\"}":{".":{},"f:lastHeartbeatTime":{},"f:lastTransitionTime":{},"f:message":{},"f:reason":{},"f:status":{},"f:type":{}},"k:{\"type\":\"FrequentUnregisterNetDevice\"}":{".":{},"f:lastHeartbeatTime":{},"f:lastTransitionTime":{},"f:message":{},"f:reason":{},"f:status":{},"f:type":{}},"k:{\"type\":\"KernelDeadlock\"}":{".":{},"f:lastHeartbeatTime":{},"f:lastTransitionTime":{},"f:message":{},"f:reason":{},"f:status":{},"f:type":{}},"k:{\"type\":\"ReadonlyFilesystem\"}":{".":{},"f:lastHeartbeatTime":{},"f:lastTransitionTime":{},"f:message":{},"f:reason":{},"f:status":{},"f:type":{}}}}} status}]},Spec:NodeSpec{PodCIDR:10.64.0.0/24,DoNotUseExternalID:,ProviderID:gce://k8s-jkns-e2e-gce-ci-reboot/us-west1-b/bootstrap-e2e-minion-group-rwsn,Unschedulable:false,Taints:[]Taint{},ConfigSource:nil,PodCIDRs:[10.64.0.0/24],},Status:NodeStatus{Capacity:ResourceList{attachable-volumes-gce-pd: {{127 0} {<nil>} 127 DecimalSI},cpu: {{2 0} {<nil>} 2 DecimalSI},ephemeral-storage: {{101203873792 0} {<nil>} 98831908Ki BinarySI},hugepages-1Gi: {{0 0} {<nil>} 0 DecimalSI},hugepages-2Mi: {{0 0} {<nil>} 0 DecimalSI},memory: {{7815438336 0} {<nil>} BinarySI},pods: {{110 0} {<nil>} 110 DecimalSI},},Allocatable:ResourceList{attachable-volumes-gce-pd: {{127 0} {<nil>} 127 DecimalSI},cpu: {{2 0} {<nil>} 2 DecimalSI},ephemeral-storage: {{91083486262 0} {<nil>} 91083486262 DecimalSI},hugepages-1Gi: {{0 0} {<nil>} 0 DecimalSI},hugepages-2Mi: {{0 0} {<nil>} 0 DecimalSI},memory: {{7553294336 0} {<nil>} BinarySI},pods: {{110 0} {<nil>} 110 DecimalSI},},Phase:,Conditions:[]NodeCondition{NodeCondition{Type:CorruptDockerOverlay2,Status:False,LastHeartbeatTime:2022-11-26 05:35:22 +0000 UTC,LastTransitionTime:2022-11-26 05:05:18 +0000 UTC,Reason:NoCorruptDockerOverlay2,Message:docker overlay2 is functioning properly,},NodeCondition{Type:FrequentUnregisterNetDevice,Status:False,LastHeartbeatTime:2022-11-26 05:35:22 +0000 UTC,LastTransitionTime:2022-11-26 05:05:18 +0000 UTC,Reason:NoFrequentUnregisterNetDevice,Message:node is functioning properly,},NodeCondition{Type:FrequentKubeletRestart,Status:False,LastHeartbeatTime:2022-11-26 05:35:22 +0000 UTC,LastTransitionTime:2022-11-26 05:05:18 +0000 UTC,Reason:NoFrequentKubeletRestart,Message:kubelet is functioning properly,},NodeCondition{Type:FrequentDockerRestart,Status:False,LastHeartbeatTime:2022-11-26 05:35:22 +0000 UTC,LastTransitionTime:2022-11-26 05:05:18 +0000 UTC,Reason:NoFrequentDockerRestart,Message:docker is functioning properly,},NodeCondition{Type:FrequentContainerdRestart,Status:False,LastHeartbeatTime:2022-11-26 05:35:22 +0000 UTC,LastTransitionTime:2022-11-26 05:05:18 +0000 UTC,Reason:NoFrequentContainerdRestart,Message:containerd is functioning properly,},NodeCondition{Type:KernelDeadlock,Status:False,LastHeartbeatTime:2022-11-26 05:35:22 +0000 UTC,LastTransitionTime:2022-11-26 05:05:18 +0000 UTC,Reason:KernelHasNoDeadlock,Message:kernel has no deadlock,},NodeCondition{Type:ReadonlyFilesystem,Status:False,LastHeartbeatTime:2022-11-26 05:35:22 +0000 UTC,LastTransitionTime:2022-11-26 05:05:18 +0000 UTC,Reason:FilesystemIsNotReadOnly,Message:Filesystem is not read-only,},NodeCondition{Type:NetworkUnavailable,Status:False,LastHeartbeatTime:2022-11-26 05:05:25 +0000 UTC,LastTransitionTime:2022-11-26 05:05:25 +0000 UTC,Reason:RouteCreated,Message:RouteController created a route,},NodeCondition{Type:MemoryPressure,Status:False,LastHeartbeatTime:2022-11-26 05:32:57 +0000 UTC,LastTransitionTime:2022-11-26 05:05:15 +0000 UTC,Reason:KubeletHasSufficientMemory,Message:kubelet has sufficient memory available,},NodeCondition{Type:DiskPressure,Status:False,LastHeartbeatTime:2022-11-26 05:32:57 +0000 UTC,LastTransitionTime:2022-11-26 05:05:15 +0000 UTC,Reason:KubeletHasNoDiskPressure,Message:kubelet has no disk pressure,},NodeCondition{Type:PIDPressure,Status:False,LastHeartbeatTime:2022-11-26 05:32:57 +0000 UTC,LastTransitionTime:2022-11-26 05:05:15 +0000 UTC,Reason:KubeletHasSufficientPID,Message:kubelet has sufficient PID available,},NodeCondition{Type:Ready,Status:True,LastHeartbeatTime:2022-11-26 05:32:57 +0000 UTC,LastTransitionTime:2022-11-26 05:05:17 +0000 UTC,Reason:KubeletReady,Message:kubelet is posting ready status. AppArmor enabled,},},Addresses:[]NodeAddress{NodeAddress{Type:InternalIP,Address:10.138.0.4,},NodeAddress{Type:ExternalIP,Address:35.230.27.45,},NodeAddress{Type:InternalDNS,Address:bootstrap-e2e-minion-group-rwsn.c.k8s-jkns-e2e-gce-ci-reboot.internal,},NodeAddress{Type:Hostname,Address:bootstrap-e2e-minion-group-rwsn.c.k8s-jkns-e2e-gce-ci-reboot.internal,},},DaemonEndpoints:NodeDaemonEndpoints{KubeletEndpoint:DaemonEndpoint{Port:10250,},},NodeInfo:NodeSystemInfo{MachineID:69e22ce1281686d6cde1b3211b34b8af,SystemUUID:69e22ce1-2816-86d6-cde1-b3211b34b8af,BootID:79e9e1fd-2a39-42ec-b72b-72836d213862,KernelVersion:5.10.123+,OSImage:Container-Optimized OS from Google,ContainerRuntimeVersion:containerd://1.7.0-beta.0-149-gd06318622,KubeletVersion:v1.27.0-alpha.0.50+70617042976dc1,KubeProxyVersion:v1.27.0-alpha.0.50+70617042976dc1,OperatingSystem:linux,Architecture:amd64,},Images:[]ContainerImage{ContainerImage{Names:[registry.k8s.io/e2e-test-images/jessie-dnsutils@sha256:24aaf2626d6b27864c29de2097e8bbb840b3a414271bf7c8995e431e47d8408e registry.k8s.io/e2e-test-images/jessie-dnsutils:1.7],SizeBytes:112030336,},ContainerImage{Names:[registry.k8s.io/e2e-test-images/volume/nfs@sha256:3bda73f2428522b0e342af80a0b9679e8594c2126f2b3cca39ed787589741b9e registry.k8s.io/e2e-test-images/volume/nfs:1.3],SizeBytes:95836203,},ContainerImage{Names:[registry.k8s.io/sig-storage/nfs-provisioner@sha256:e943bb77c7df05ebdc8c7888b2db289b13bf9f012d6a3a5a74f14d4d5743d439 registry.k8s.io/sig-storage/nfs-provisioner:v3.0.1],SizeBytes:90632047,},ContainerImage{Names:[registry.k8s.io/kube-proxy-amd64:v1.27.0-alpha.0.50_70617042976dc1],SizeBytes:67201736,},ContainerImage{Names:[registry.k8s.io/e2e-test-images/agnhost@sha256:16bbf38c463a4223d8cfe4da12bc61010b082a79b4bb003e2d3ba3ece5dd5f9e registry.k8s.io/e2e-test-images/agnhost:2.43],SizeBytes:51706353,},ContainerImage{Names:[gke.gcr.io/prometheus-to-sd@sha256:e739643c3939ba0b161425f45a1989eedfc4a3b166db9a7100863296b4c70510 gke.gcr.io/prometheus-to-sd:v0.11.1-gke.1],SizeBytes:48742566,},ContainerImage{Names:[registry.k8s.io/metrics-server/metrics-server@sha256:6385aec64bb97040a5e692947107b81e178555c7a5b71caa90d733e4130efc10 registry.k8s.io/metrics-server/metrics-server:v0.5.2],SizeBytes:26023008,},ContainerImage{Names:[registry.k8s.io/sig-storage/csi-provisioner@sha256:ee3b525d5b89db99da3b8eb521d9cd90cb6e9ef0fbb651e98bb37be78d36b5b8 registry.k8s.io/sig-storage/csi-provisioner:v3.3.0],SizeBytes:25491225,},ContainerImage{Names:[registry.k8s.io/sig-storage/csi-resizer@sha256:425d8f1b769398127767b06ed97ce62578a3179bcb99809ce93a1649e025ffe7 registry.k8s.io/sig-storage/csi-resizer:v1.6.0],SizeBytes:24148884,},ContainerImage{Names:[registry.k8s.io/sig-storage/csi-snapshotter@sha256:291334908ddf71a4661fd7f6d9d97274de8a5378a2b6fdfeb2ce73414a34f82f registry.k8s.io/sig-storage/csi-snapshotter:v6.1.0],SizeBytes:23881995,},ContainerImage{Names:[registry.k8s.io/sig-storage/csi-attacher@sha256:9a685020911e2725ad019dbce6e4a5ab93d51e3d4557f115e64343345e05781b registry.k8s.io/sig-storage/csi-attacher:v4.0.0],SizeBytes:23847201,},ContainerImage{Names:[registry.k8s.io/sig-storage/snapshot-controller@sha256:823c75d0c45d1427f6d850070956d9ca657140a7bbf828381541d1d808475280 registry.k8s.io/sig-storage/snapshot-controller:v6.1.0],SizeBytes:22620891,},ContainerImage{Names:[registry.k8s.io/sig-storage/hostpathplugin@sha256:92257881c1d6493cf18299a24af42330f891166560047902b8d431fb66b01af5 registry.k8s.io/sig-storage/hostpathplugin:v1.9.0],SizeBytes:18758628,},ContainerImage{Names:[registry.k8s.io/cpa/cluster-proportional-autoscaler@sha256:fd636b33485c7826fb20ef0688a83ee0910317dbb6c0c6f3ad14661c1db25def registry.k8s.io/cpa/cluster-proportional-autoscaler:1.8.4],SizeBytes:15209393,},ContainerImage{Names:[registry.k8s.io/coredns/coredns@sha256:8e352a029d304ca7431c6507b56800636c321cb52289686a581ab70aaa8a2e2a registry.k8s.io/coredns/coredns:v1.9.3],SizeBytes:14837849,},ContainerImage{Names:[registry.k8s.io/autoscaling/addon-resizer@sha256:43f129b81d28f0fdd54de6d8e7eacd5728030782e03db16087fc241ad747d3d6 registry.k8s.io/autoscaling/addon-resizer:1.8.14],SizeBytes:10153852,},ContainerImage{Names:[registry.k8s.io/sig-storage/csi-node-driver-registrar@sha256:0103eee7c35e3e0b5cd8cdca9850dc71c793cdeb6669d8be7a89440da2d06ae4 registry.k8s.io/sig-storage/csi-node-driver-registrar:v2.5.1],SizeBytes:9133109,},ContainerImage{Names:[registry.k8s.io/sig-storage/livenessprobe@sha256:933940f13b3ea0abc62e656c1aa5c5b47c04b15d71250413a6b821bd0c58b94e registry.k8s.io/sig-storage/livenessprobe:v2.7.0],SizeBytes:8688564,},ContainerImage{Names:[registry.k8s.io/kas-network-proxy/proxy-agent@sha256:48f2a4ec3e10553a81b8dd1c6fa5fe4bcc9617f78e71c1ca89c6921335e2d7da registry.k8s.io/kas-network-proxy/proxy-agent:v0.0.33],SizeBytes:8512162,},ContainerImage{Names:[registry.k8s.io/networking/ingress-gce-404-server-with-metrics-amd64@sha256:7eb7b3cee4d33c10c49893ad3c386232b86d4067de5251294d4c620d6e072b93 registry.k8s.io/networking/ingress-gce-404-server-with-metrics-amd64:v1.10.11],SizeBytes:6463068,},ContainerImage{Names:[registry.k8s.io/metadata-proxy@sha256:e914645f22e946bce5165737e1b244e0a296ad1f0f81a9531adc57af2780978a registry.k8s.io/metadata-proxy:v0.1.12],SizeBytes:5301657,},ContainerImage{Names:[registry.k8s.io/e2e-test-images/busybox@sha256:c318242786b139d18676b1c09a0ad7f15fc17f8f16a5b2e625cd0dc8c9703daf registry.k8s.io/e2e-test-images/busybox:1.29-2],SizeBytes:732424,},ContainerImage{Names:[registry.k8s.io/e2e-test-images/busybox@sha256:2e0f836850e09b8b7cc937681d6194537a09fbd5f6b9e08f4d646a85128e8937 registry.k8s.io/e2e-test-images/busybox:1.29-4],SizeBytes:731990,},ContainerImage{Names:[registry.k8s.io/pause@sha256:7031c1b283388d2c2e09b57badb803c05ebed362dc88d84b480cc47f72a21097 registry.k8s.io/pause:3.9],SizeBytes:321520,},ContainerImage{Names:[registry.k8s.io/pause@sha256:9001185023633d17a2f98ff69b6ff2615b8ea02a825adffa40422f51dfdcde9d registry.k8s.io/pause:3.8],SizeBytes:311286,},},VolumesInUse:[kubernetes.io/csi/csi-hostpath-multivolume-4683^6fe7a3e3-6d49-11ed-ac34-8a6e303c97ed kubernetes.io/csi/csi-hostpath-multivolume-4683^7ad9a9cd-6d49-11ed-b2a0-f2f9c395d3b2 kubernetes.io/csi/csi-hostpath-multivolume-6872^8dfb6ea6-6d49-11ed-bbbd-7a3847645704 kubernetes.io/csi/csi-hostpath-provisioning-6984^3b0457df-6d48-11ed-9e4d-d626cdd9a50a],VolumesAttached:[]AttachedVolume{AttachedVolume{Name:kubernetes.io/csi/csi-hostpath-multivolume-6872^8dfb6ea6-6d49-11ed-bbbd-7a3847645704,DevicePath:,},AttachedVolume{Name:kubernetes.io/csi/csi-hostpath-provisioning-6984^3b0457df-6d48-11ed-9e4d-d626cdd9a50a,DevicePath:,},AttachedVolume{Name:kubernetes.io/csi/csi-hostpath-multivolume-4683^7ad9a9cd-6d49-11ed-b2a0-f2f9c395d3b2,DevicePath:,},AttachedVolume{Name:kubernetes.io/csi/csi-hostpath-multivolume-4683^6fe7a3e3-6d49-11ed-ac34-8a6e303c97ed,DevicePath:,},},Config:nil,},} Nov 26 05:35:25.211: INFO: Logging kubelet events for node bootstrap-e2e-minion-group-rwsn Nov 26 05:35:25.255: INFO: Logging pods the kubelet thinks is on node bootstrap-e2e-minion-group-rwsn Nov 26 05:35:25.297: INFO: Unable to retrieve kubelet pods for node bootstrap-e2e-minion-group-rwsn: error trying to reach service: No agent available Nov 26 05:35:25.297: INFO: Logging node info for node bootstrap-e2e-minion-group-vhdj Nov 26 05:35:25.339: INFO: Node Info: &Node{ObjectMeta:{bootstrap-e2e-minion-group-vhdj 0d27fd16-9ec8-4428-bb1e-97fead6c7a66 12328 0 2022-11-26 05:05:23 +0000 UTC <nil> <nil> map[beta.kubernetes.io/arch:amd64 beta.kubernetes.io/instance-type:n1-standard-2 beta.kubernetes.io/os:linux cloud.google.com/metadata-proxy-ready:true failure-domain.beta.kubernetes.io/region:us-west1 failure-domain.beta.kubernetes.io/zone:us-west1-b kubernetes.io/arch:amd64 kubernetes.io/hostname:bootstrap-e2e-minion-group-vhdj kubernetes.io/os:linux node.kubernetes.io/instance-type:n1-standard-2 topology.hostpath.csi/node:bootstrap-e2e-minion-group-vhdj topology.kubernetes.io/region:us-west1 topology.kubernetes.io/zone:us-west1-b] map[csi.volume.kubernetes.io/nodeid:{"csi-hostpath-volumeio-4786":"bootstrap-e2e-minion-group-vhdj","csi-mock-csi-mock-volumes-5516":"bootstrap-e2e-minion-group-vhdj"} node.alpha.kubernetes.io/ttl:0 volumes.kubernetes.io/controller-managed-attach-detach:true] [] [] [{kubelet Update v1 2022-11-26 05:05:23 +0000 UTC FieldsV1 {"f:metadata":{"f:annotations":{".":{},"f:volumes.kubernetes.io/controller-managed-attach-detach":{}},"f:labels":{".":{},"f:beta.kubernetes.io/arch":{},"f:beta.kubernetes.io/instance-type":{},"f:beta.kubernetes.io/os":{},"f:cloud.google.com/metadata-proxy-ready":{},"f:failure-domain.beta.kubernetes.io/region":{},"f:failure-domain.beta.kubernetes.io/zone":{},"f:kubernetes.io/arch":{},"f:kubernetes.io/hostname":{},"f:kubernetes.io/os":{},"f:node.kubernetes.io/instance-type":{},"f:topology.kubernetes.io/region":{},"f:topology.kubernetes.io/zone":{}}},"f:spec":{"f:providerID":{}}} } {kube-controller-manager Update v1 2022-11-26 05:05:24 +0000 UTC FieldsV1 {"f:metadata":{"f:annotations":{"f:node.alpha.kubernetes.io/ttl":{}}},"f:spec":{"f:podCIDR":{},"f:podCIDRs":{".":{},"v:\"10.64.3.0/24\"":{}}}} } {kube-controller-manager Update v1 2022-11-26 05:17:38 +0000 UTC FieldsV1 {"f:status":{"f:conditions":{"k:{\"type\":\"NetworkUnavailable\"}":{"f:lastHeartbeatTime":{},"f:lastTransitionTime":{},"f:message":{},"f:reason":{},"f:status":{}}}}} status} {node-problem-detector Update v1 2022-11-26 05:30:29 +0000 UTC FieldsV1 {"f:status":{"f:conditions":{"k:{\"type\":\"CorruptDockerOverlay2\"}":{".":{},"f:lastHeartbeatTime":{},"f:lastTransitionTime":{},"f:message":{},"f:reason":{},"f:status":{},"f:type":{}},"k:{\"type\":\"FrequentContainerdRestart\"}":{".":{},"f:lastHeartbeatTime":{},"f:lastTransitionTime":{},"f:message":{},"f:reason":{},"f:status":{},"f:type":{}},"k:{\"type\":\"FrequentDockerRestart\"}":{".":{},"f:lastHeartbeatTime":{},"f:lastTransitionTime":{},"f:message":{},"f:reason":{},"f:status":{},"f:type":{}},"k:{\"type\":\"FrequentKubeletRestart\"}":{".":{},"f:lastHeartbeatTime":{},"f:lastTransitionTime":{},"f:message":{},"f:reason":{},"f:status":{},"f:type":{}},"k:{\"type\":\"FrequentUnregisterNetDevice\"}":{".":{},"f:lastHeartbeatTime":{},"f:lastTransitionTime":{},"f:message":{},"f:reason":{},"f:status":{},"f:type":{}},"k:{\"type\":\"KernelDeadlock\"}":{".":{},"f:lastHeartbeatTime":{},"f:lastTransitionTime":{},"f:message":{},"f:reason":{},"f:status":{},"f:type":{}},"k:{\"type\":\"ReadonlyFilesystem\"}":{".":{},"f:lastHeartbeatTime":{},"f:lastTransitionTime":{},"f:message":{},"f:reason":{},"f:status":{},"f:type":{}}}}} status} {kubelet Update v1 2022-11-26 05:35:12 +0000 UTC FieldsV1 {"f:metadata":{"f:annotations":{"f:csi.volume.kubernetes.io/nodeid":{}},"f:labels":{"f:topology.hostpath.csi/node":{}}},"f:status":{"f:conditions":{"k:{\"type\":\"DiskPressure\"}":{"f:lastHeartbeatTime":{}},"k:{\"type\":\"MemoryPressure\"}":{"f:lastHeartbeatTime":{}},"k:{\"type\":\"PIDPressure\"}":{"f:lastHeartbeatTime":{}},"k:{\"type\":\"Ready\"}":{"f:lastHeartbeatTime":{},"f:lastTransitionTime":{},"f:message":{},"f:reason":{},"f:status":{}}},"f:images":{}}} status}]},Spec:NodeSpec{PodCIDR:10.64.3.0/24,DoNotUseExternalID:,ProviderID:gce://k8s-jkns-e2e-gce-ci-reboot/us-west1-b/bootstrap-e2e-minion-group-vhdj,Unschedulable:false,Taints:[]Taint{},ConfigSource:nil,PodCIDRs:[10.64.3.0/24],},Status:NodeStatus{Capacity:ResourceList{attachable-volumes-gce-pd: {{127 0} {<nil>} 127 DecimalSI},cpu: {{2 0} {<nil>} 2 DecimalSI},ephemeral-storage: {{101203873792 0} {<nil>} 98831908Ki BinarySI},hugepages-1Gi: {{0 0} {<nil>} 0 DecimalSI},hugepages-2Mi: {{0 0} {<nil>} 0 DecimalSI},memory: {{7815438336 0} {<nil>} BinarySI},pods: {{110 0} {<nil>} 110 DecimalSI},},Allocatable:ResourceList{attachable-volumes-gce-pd: {{127 0} {<nil>} 127 DecimalSI},cpu: {{2 0} {<nil>} 2 DecimalSI},ephemeral-storage: {{91083486262 0} {<nil>} 91083486262 DecimalSI},hugepages-1Gi: {{0 0} {<nil>} 0 DecimalSI},hugepages-2Mi: {{0 0} {<nil>} 0 DecimalSI},memory: {{7553294336 0} {<nil>} BinarySI},pods: {{110 0} {<nil>} 110 DecimalSI},},Phase:,Conditions:[]NodeCondition{NodeCondition{Type:KernelDeadlock,Status:False,LastHeartbeatTime:2022-11-26 05:30:29 +0000 UTC,LastTransitionTime:2022-11-26 05:05:26 +0000 UTC,Reason:KernelHasNoDeadlock,Message:kernel has no deadlock,},NodeCondition{Type:ReadonlyFilesystem,Status:False,LastHeartbeatTime:2022-11-26 05:30:29 +0000 UTC,LastTransitionTime:2022-11-26 05:05:26 +0000 UTC,Reason:FilesystemIsNotReadOnly,Message:Filesystem is not read-only,},NodeCondition{Type:CorruptDockerOverlay2,Status:False,LastHeartbeatTime:2022-11-26 05:30:29 +0000 UTC,LastTransitionTime:2022-11-26 05:05:26 +0000 UTC,Reason:NoCorruptDockerOverlay2,Message:docker overlay2 is functioning properly,},NodeCondition{Type:FrequentUnregisterNetDevice,Status:False,LastHeartbeatTime:2022-11-26 05:30:29 +0000 UTC,LastTransitionTime:2022-11-26 05:05:26 +0000 UTC,Reason:NoFrequentUnregisterNetDevice,Message:node is functioning properly,},NodeCondition{Type:FrequentKubeletRestart,Status:False,LastHeartbeatTime:2022-11-26 05:30:29 +0000 UTC,LastTransitionTime:2022-11-26 05:05:26 +0000 UTC,Reason:NoFrequentKubeletRestart,Message:kubelet is functioning properly,},NodeCondition{Type:FrequentDockerRestart,Status:False,LastHeartbeatTime:2022-11-26 05:30:29 +0000 UTC,LastTransitionTime:2022-11-26 05:05:26 +0000 UTC,Reason:NoFrequentDockerRestart,Message:docker is functioning properly,},NodeCondition{Type:FrequentContainerdRestart,Status:False,LastHeartbeatTime:2022-11-26 05:30:29 +0000 UTC,LastTransitionTime:2022-11-26 05:05:26 +0000 UTC,Reason:NoFrequentContainerdRestart,Message:containerd is functioning properly,},NodeCondition{Type:NetworkUnavailable,Status:False,LastHeartbeatTime:2022-11-26 05:05:39 +0000 UTC,LastTransitionTime:2022-11-26 05:05:39 +0000 UTC,Reason:RouteCreated,Message:RouteController created a route,},NodeCondition{Type:MemoryPressure,Status:False,LastHeartbeatTime:2022-11-26 05:32:57 +0000 UTC,LastTransitionTime:2022-11-26 05:05:23 +0000 UTC,Reason:KubeletHasSufficientMemory,Message:kubelet has sufficient memory available,},NodeCondition{Type:DiskPressure,Status:False,LastHeartbeatTime:2022-11-26 05:32:57 +0000 UTC,LastTransitionTime:2022-11-26 05:05:23 +0000 UTC,Reason:KubeletHasNoDiskPressure,Message:kubelet has no disk pressure,},NodeCondition{Type:PIDPressure,Status:False,LastHeartbeatTime:2022-11-26 05:32:57 +0000 UTC,LastTransitionTime:2022-11-26 05:05:23 +0000 UTC,Reason:KubeletHasSufficientPID,Message:kubelet has sufficient PID available,},NodeCondition{Type:Ready,Status:True,LastHeartbeatTime:2022-11-26 05:32:57 +0000 UTC,LastTransitionTime:2022-11-26 05:05:24 +0000 UTC,Reason:KubeletReady,Message:kubelet is posting ready status. AppArmor enabled,},},Addresses:[]NodeAddress{NodeAddress{Type:InternalIP,Address:10.138.0.3,},NodeAddress{Type:ExternalIP,Address:34.83.94.215,},NodeAddress{Type:InternalDNS,Address:bootstrap-e2e-minion-group-vhdj.c.k8s-jkns-e2e-gce-ci-reboot.internal,},NodeAddress{Type:Hostname,Address:bootstrap-e2e-minion-group-vhdj.c.k8s-jkns-e2e-gce-ci-reboot.internal,},},DaemonEndpoints:NodeDaemonEndpoints{KubeletEndpoint:DaemonEndpoint{Port:10250,},},NodeInfo:NodeSystemInfo{MachineID:31b0953ad3e740aeb7fc1a89edb5435a,SystemUUID:31b0953a-d3e7-40ae-b7fc-1a89edb5435a,BootID:14c054d0-bda9-4ddf-b9f6-e1742be33d25,KernelVersion:5.10.123+,OSImage:Container-Optimized OS from Google,ContainerRuntimeVersion:containerd://1.7.0-beta.0-149-gd06318622,KubeletVersion:v1.27.0-alpha.0.50+70617042976dc1,KubeProxyVersion:v1.27.0-alpha.0.50+70617042976dc1,OperatingSystem:linux,Architecture:amd64,},Images:[]ContainerImage{ContainerImage{Names:[registry.k8s.io/e2e-test-images/jessie-dnsutils@sha256:24aaf2626d6b27864c29de2097e8bbb840b3a414271bf7c8995e431e47d8408e registry.k8s.io/e2e-test-images/jessie-dnsutils:1.7],SizeBytes:112030336,},ContainerImage{Names:[registry.k8s.io/sig-storage/nfs-provisioner@sha256:e943bb77c7df05ebdc8c7888b2db289b13bf9f012d6a3a5a74f14d4d5743d439 registry.k8s.io/sig-storage/nfs-provisioner:v3.0.1],SizeBytes:90632047,},ContainerImage{Names:[registry.k8s.io/kube-proxy-amd64:v1.27.0-alpha.0.50_70617042976dc1],SizeBytes:67201736,},ContainerImage{Names:[registry.k8s.io/e2e-test-images/agnhost@sha256:16bbf38c463a4223d8cfe4da12bc61010b082a79b4bb003e2d3ba3ece5dd5f9e registry.k8s.io/e2e-test-images/agnhost:2.43],SizeBytes:51706353,},ContainerImage{Names:[gke.gcr.io/prometheus-to-sd@sha256:e739643c3939ba0b161425f45a1989eedfc4a3b166db9a7100863296b4c70510 gke.gcr.io/prometheus-to-sd:v0.11.1-gke.1],SizeBytes:48742566,},ContainerImage{Names:[registry.k8s.io/e2e-test-images/httpd@sha256:148b022f5c5da426fc2f3c14b5c0867e58ef05961510c84749ac1fddcb0fef22 registry.k8s.io/e2e-test-images/httpd:2.4.38-4],SizeBytes:40764257,},ContainerImage{Names:[registry.k8s.io/sig-storage/csi-provisioner@sha256:ee3b525d5b89db99da3b8eb521d9cd90cb6e9ef0fbb651e98bb37be78d36b5b8 registry.k8s.io/sig-storage/csi-provisioner:v3.3.0],SizeBytes:25491225,},ContainerImage{Names:[registry.k8s.io/sig-storage/csi-resizer@sha256:425d8f1b769398127767b06ed97ce62578a3179bcb99809ce93a1649e025ffe7 registry.k8s.io/sig-storage/csi-resizer:v1.6.0],SizeBytes:24148884,},ContainerImage{Names:[registry.k8s.io/sig-storage/csi-snapshotter@sha256:291334908ddf71a4661fd7f6d9d97274de8a5378a2b6fdfeb2ce73414a34f82f registry.k8s.io/sig-storage/csi-snapshotter:v6.1.0],SizeBytes:23881995,},ContainerImage{Names:[registry.k8s.io/sig-storage/csi-attacher@sha256:9a685020911e2725ad019dbce6e4a5ab93d51e3d4557f115e64343345e05781b registry.k8s.io/sig-storage/csi-attacher:v4.0.0],SizeBytes:23847201,},ContainerImage{Names:[registry.k8s.io/sig-storage/hostpathplugin@sha256:92257881c1d6493cf18299a24af42330f891166560047902b8d431fb66b01af5 registry.k8s.io/sig-storage/hostpathplugin:v1.9.0],SizeBytes:18758628,},ContainerImage{Names:[registry.k8s.io/sig-storage/csi-node-driver-registrar@sha256:0103eee7c35e3e0b5cd8cdca9850dc71c793cdeb6669d8be7a89440da2d06ae4 registry.k8s.io/sig-storage/csi-node-driver-registrar:v2.5.1],SizeBytes:9133109,},ContainerImage{Names:[registry.k8s.io/sig-storage/livenessprobe@sha256:933940f13b3ea0abc62e656c1aa5c5b47c04b15d71250413a6b821bd0c58b94e registry.k8s.io/sig-storage/livenessprobe:v2.7.0],SizeBytes:8688564,},ContainerImage{Names:[registry.k8s.io/kas-network-proxy/proxy-agent@sha256:48f2a4ec3e10553a81b8dd1c6fa5fe4bcc9617f78e71c1ca89c6921335e2d7da registry.k8s.io/kas-network-proxy/proxy-agent:v0.0.33],SizeBytes:8512162,},ContainerImage{Names:[registry.k8s.io/metadata-proxy@sha256:e914645f22e946bce5165737e1b244e0a296ad1f0f81a9531adc57af2780978a registry.k8s.io/metadata-proxy:v0.1.12],SizeBytes:5301657,},ContainerImage{Names:[registry.k8s.io/e2e-test-images/busybox@sha256:2e0f836850e09b8b7cc937681d6194537a09fbd5f6b9e08f4d646a85128e8937 registry.k8s.io/e2e-test-images/busybox:1.29-4],SizeBytes:731990,},ContainerImage{Names:[registry.k8s.io/pause@sha256:9001185023633d17a2f98ff69b6ff2615b8ea02a825adffa40422f51dfdcde9d registry.k8s.io/pause:3.8],SizeBytes:311286,},},VolumesInUse:[],VolumesAttached:[]AttachedVolume{},Config:nil,},} Nov 26 05:35:25.339: INFO: Logging kubelet events for node bootstrap-e2e-minion-group-vhdj Nov 26 05:35:25.382: INFO: Logging pods the kubelet thinks is on node bootstrap-e2e-minion-group-vhdj Nov 26 05:35:25.425: INFO: Unable to retrieve kubelet pods for node bootstrap-e2e-minion-group-vhdj: error trying to reach service: No agent available [DeferCleanup (Each)] [sig-api-machinery] Servers with support for API chunking tear down framework | framework.go:193 STEP: Destroying namespace "chunking-6807" for this suite. 11/26/22 05:35:25.425
Filter through log files | View test history on testgrid
go run hack/e2e.go -v --test --test_args='--ginkgo.focus=Kubernetes\se2e\ssuite\s\[It\]\s\[sig\-apps\]\sCronJob\sshould\snot\sschedule\snew\sjobs\swhen\sForbidConcurrent\s\[Slow\]\s\[Conformance\]$'
test/e2e/apps/cronjob.go:152 k8s.io/kubernetes/test/e2e/apps.glob..func2.3() test/e2e/apps/cronjob.go:152 +0xa3c There were additional failures detected after the initial failure: [FAILED] Nov 26 05:13:57.974: failed to list events in namespace "cronjob-6006": Get "https://35.227.179.144/api/v1/namespaces/cronjob-6006/events": dial tcp 35.227.179.144:443: connect: connection refused In [DeferCleanup (Each)] at: test/e2e/framework/debug/dump.go:44 ---------- [FAILED] Nov 26 05:13:58.014: Couldn't delete ns: "cronjob-6006": Delete "https://35.227.179.144/api/v1/namespaces/cronjob-6006": dial tcp 35.227.179.144:443: connect: connection refused (&url.Error{Op:"Delete", URL:"https://35.227.179.144/api/v1/namespaces/cronjob-6006", Err:(*net.OpError)(0xc00107f090)}) In [DeferCleanup (Each)] at: test/e2e/framework/framework.go:370from junit_01.xml
[BeforeEach] [sig-apps] CronJob set up framework | framework.go:178 STEP: Creating a kubernetes client 11/26/22 05:08:35.204 Nov 26 05:08:35.204: INFO: >>> kubeConfig: /workspace/.kube/config STEP: Building a namespace api object, basename cronjob 11/26/22 05:08:35.205 STEP: Waiting for a default service account to be provisioned in namespace 11/26/22 05:08:35.438 STEP: Waiting for kube-root-ca.crt to be provisioned in namespace 11/26/22 05:08:35.53 [BeforeEach] [sig-apps] CronJob test/e2e/framework/metrics/init/init.go:31 [It] should not schedule new jobs when ForbidConcurrent [Slow] [Conformance] test/e2e/apps/cronjob.go:124 STEP: Creating a ForbidConcurrent cronjob 11/26/22 05:08:35.617 STEP: Ensuring a job is scheduled 11/26/22 05:08:35.689 STEP: Ensuring exactly one is scheduled 11/26/22 05:09:43.733 STEP: Ensuring exactly one running job exists by listing jobs explicitly 11/26/22 05:09:43.774 STEP: Ensuring no more jobs are scheduled 11/26/22 05:09:43.815 ------------------------------ Progress Report for Ginkgo Process #10 Automatically polling progress: [sig-apps] CronJob should not schedule new jobs when ForbidConcurrent [Slow] [Conformance] (Spec Runtime: 5m0.414s) test/e2e/apps/cronjob.go:124 In [It] (Node Runtime: 5m0s) test/e2e/apps/cronjob.go:124 At [By Step] Ensuring no more jobs are scheduled (Step Runtime: 3m51.803s) test/e2e/apps/cronjob.go:146 Spec Goroutine goroutine 738 [select] k8s.io/kubernetes/vendor/k8s.io/apimachinery/pkg/util/wait.WaitForWithContext({0x7fe0bc8, 0xc0000820c8}, 0xc0024c3d70, 0x2fdb16a?) vendor/k8s.io/apimachinery/pkg/util/wait/wait.go:660 k8s.io/kubernetes/vendor/k8s.io/apimachinery/pkg/util/wait.poll({0x7fe0bc8, 0xc0000820c8}, 0x10?, 0x2fd9d05?, 0x30?) vendor/k8s.io/apimachinery/pkg/util/wait/wait.go:596 k8s.io/kubernetes/vendor/k8s.io/apimachinery/pkg/util/wait.PollWithContext({0x7fe0bc8, 0xc0000820c8}, 0x766e270?, 0xc000cd9d60?, 0x262a967?) vendor/k8s.io/apimachinery/pkg/util/wait/wait.go:460 k8s.io/kubernetes/vendor/k8s.io/apimachinery/pkg/util/wait.Poll(0x0?, 0xc0d885c5f0970263?, 0x2f8d9aeb15?) vendor/k8s.io/apimachinery/pkg/util/wait/wait.go:445 > k8s.io/kubernetes/test/e2e/apps.waitForActiveJobs({0x801de88?, 0xc002ae4680}, {0xc0029afaf0, 0xc}, {0xc0011d6120, 0x6}, 0x2) test/e2e/apps/cronjob.go:593 > k8s.io/kubernetes/test/e2e/apps.glob..func2.3() test/e2e/apps/cronjob.go:147 k8s.io/kubernetes/vendor/github.com/onsi/ginkgo/v2/internal.extractBodyFunction.func3({0xc00101abd0, 0xc001cc2420}) vendor/github.com/onsi/ginkgo/v2/internal/node.go:449 k8s.io/kubernetes/vendor/github.com/onsi/ginkgo/v2/internal.(*Suite).runNode.func2() vendor/github.com/onsi/ginkgo/v2/internal/suite.go:750 k8s.io/kubernetes/vendor/github.com/onsi/ginkgo/v2/internal.(*Suite).runNode vendor/github.com/onsi/ginkgo/v2/internal/suite.go:738 ------------------------------ ------------------------------ Progress Report for Ginkgo Process #10 Automatically polling progress: [sig-apps] CronJob should not schedule new jobs when ForbidConcurrent [Slow] [Conformance] (Spec Runtime: 5m20.415s) test/e2e/apps/cronjob.go:124 In [It] (Node Runtime: 5m20.002s) test/e2e/apps/cronjob.go:124 At [By Step] Ensuring no more jobs are scheduled (Step Runtime: 4m11.804s) test/e2e/apps/cronjob.go:146 Spec Goroutine goroutine 738 [select] k8s.io/kubernetes/vendor/k8s.io/apimachinery/pkg/util/wait.WaitForWithContext({0x7fe0bc8, 0xc0000820c8}, 0xc0024c3d70, 0x2fdb16a?) vendor/k8s.io/apimachinery/pkg/util/wait/wait.go:660 k8s.io/kubernetes/vendor/k8s.io/apimachinery/pkg/util/wait.poll({0x7fe0bc8, 0xc0000820c8}, 0x10?, 0x2fd9d05?, 0x30?) vendor/k8s.io/apimachinery/pkg/util/wait/wait.go:596 k8s.io/kubernetes/vendor/k8s.io/apimachinery/pkg/util/wait.PollWithContext({0x7fe0bc8, 0xc0000820c8}, 0x766e270?, 0xc000cd9d60?, 0x262a967?) vendor/k8s.io/apimachinery/pkg/util/wait/wait.go:460 k8s.io/kubernetes/vendor/k8s.io/apimachinery/pkg/util/wait.Poll(0x0?, 0xc0d885c5f0970263?, 0x2f8d9aeb15?) vendor/k8s.io/apimachinery/pkg/util/wait/wait.go:445 > k8s.io/kubernetes/test/e2e/apps.waitForActiveJobs({0x801de88?, 0xc002ae4680}, {0xc0029afaf0, 0xc}, {0xc0011d6120, 0x6}, 0x2) test/e2e/apps/cronjob.go:593 > k8s.io/kubernetes/test/e2e/apps.glob..func2.3() test/e2e/apps/cronjob.go:147 k8s.io/kubernetes/vendor/github.com/onsi/ginkgo/v2/internal.extractBodyFunction.func3({0xc00101abd0, 0xc001cc2420}) vendor/github.com/onsi/ginkgo/v2/internal/node.go:449 k8s.io/kubernetes/vendor/github.com/onsi/ginkgo/v2/internal.(*Suite).runNode.func2() vendor/github.com/onsi/ginkgo/v2/internal/suite.go:750 k8s.io/kubernetes/vendor/github.com/onsi/ginkgo/v2/internal.(*Suite).runNode vendor/github.com/onsi/ginkgo/v2/internal/suite.go:738 ------------------------------ STEP: Removing cronjob 11/26/22 05:13:57.855 Nov 26 05:13:57.895: INFO: Unexpected error: Failed to delete CronJob forbid in namespace cronjob-6006: <*url.Error | 0xc003a14c90>: { Op: "Delete", URL: "https://35.227.179.144/apis/batch/v1/namespaces/cronjob-6006/cronjobs/forbid", Err: <*net.OpError | 0xc00222f5e0>{ Op: "dial", Net: "tcp", Source: nil, Addr: <*net.TCPAddr | 0xc00204c060>{ IP: [0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 255, 255, 35, 227, 179, 144], Port: 443, Zone: "", }, Err: <*os.SyscallError | 0xc00152f2c0>{ Syscall: "connect", Err: <syscall.Errno>0x6f, }, }, } Nov 26 05:13:57.895: FAIL: Failed to delete CronJob forbid in namespace cronjob-6006: Delete "https://35.227.179.144/apis/batch/v1/namespaces/cronjob-6006/cronjobs/forbid": dial tcp 35.227.179.144:443: connect: connection refused Full Stack Trace k8s.io/kubernetes/test/e2e/apps.glob..func2.3() test/e2e/apps/cronjob.go:152 +0xa3c [AfterEach] [sig-apps] CronJob test/e2e/framework/node/init/init.go:32 Nov 26 05:13:57.895: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready [DeferCleanup (Each)] [sig-apps] CronJob test/e2e/framework/metrics/init/init.go:33 [DeferCleanup (Each)] [sig-apps] CronJob dump namespaces | framework.go:196 STEP: dump namespace information after failure 11/26/22 05:13:57.935 STEP: Collecting events from namespace "cronjob-6006". 11/26/22 05:13:57.935 Nov 26 05:13:57.974: INFO: Unexpected error: failed to list events in namespace "cronjob-6006": <*url.Error | 0xc00204c090>: { Op: "Get", URL: "https://35.227.179.144/api/v1/namespaces/cronjob-6006/events", Err: <*net.OpError | 0xc00107ec30>{ Op: "dial", Net: "tcp", Source: nil, Addr: <*net.TCPAddr | 0xc0016a9e30>{ IP: [0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 255, 255, 35, 227, 179, 144], Port: 443, Zone: "", }, Err: <*os.SyscallError | 0xc0014e4660>{ Syscall: "connect", Err: <syscall.Errno>0x6f, }, }, } Nov 26 05:13:57.974: FAIL: failed to list events in namespace "cronjob-6006": Get "https://35.227.179.144/api/v1/namespaces/cronjob-6006/events": dial tcp 35.227.179.144:443: connect: connection refused Full Stack Trace k8s.io/kubernetes/test/e2e/framework/debug.dumpEventsInNamespace(0xc001cda5c0, {0xc0029afaf0, 0xc}) test/e2e/framework/debug/dump.go:44 +0x191 k8s.io/kubernetes/test/e2e/framework/debug.DumpAllNamespaceInfo({0x801de88, 0xc002ae4680}, {0xc0029afaf0, 0xc}) test/e2e/framework/debug/dump.go:62 +0x8d k8s.io/kubernetes/test/e2e/framework/debug/init.init.0.func1.1(0xc001cda650?, {0xc0029afaf0?, 0x7fa7740?}) test/e2e/framework/debug/init/init.go:34 +0x32 k8s.io/kubernetes/test/e2e/framework.(*Framework).dumpNamespaceInfo.func1() test/e2e/framework/framework.go:274 +0x6d k8s.io/kubernetes/test/e2e/framework.(*Framework).dumpNamespaceInfo(0xc0006f6c30) test/e2e/framework/framework.go:271 +0x179 reflect.Value.call({0x6627cc0?, 0xc001291a90?, 0xc001dfdfb0?}, {0x75b6e72, 0x4}, {0xae73300, 0x0, 0xc001e90a48?}) /usr/local/go/src/reflect/value.go:584 +0x8c5 reflect.Value.Call({0x6627cc0?, 0xc001291a90?, 0x29449fc?}, {0xae73300?, 0xc001dfdf80?, 0x2a6d786?}) /usr/local/go/src/reflect/value.go:368 +0xbc [DeferCleanup (Each)] [sig-apps] CronJob tear down framework | framework.go:193 STEP: Destroying namespace "cronjob-6006" for this suite. 11/26/22 05:13:57.975 Nov 26 05:13:58.014: FAIL: Couldn't delete ns: "cronjob-6006": Delete "https://35.227.179.144/api/v1/namespaces/cronjob-6006": dial tcp 35.227.179.144:443: connect: connection refused (&url.Error{Op:"Delete", URL:"https://35.227.179.144/api/v1/namespaces/cronjob-6006", Err:(*net.OpError)(0xc00107f090)}) Full Stack Trace k8s.io/kubernetes/test/e2e/framework.(*Framework).AfterEach.func1() test/e2e/framework/framework.go:370 +0x4fe k8s.io/kubernetes/test/e2e/framework.(*Framework).AfterEach(0xc0006f6c30) test/e2e/framework/framework.go:383 +0x1ca reflect.Value.call({0x6627cc0?, 0xc0012919d0?, 0xc001059fb0?}, {0x75b6e72, 0x4}, {0xae73300, 0x0, 0x0?}) /usr/local/go/src/reflect/value.go:584 +0x8c5 reflect.Value.Call({0x6627cc0?, 0xc0012919d0?, 0x0?}, {0xae73300?, 0x5?, 0xc00050cc18?}) /usr/local/go/src/reflect/value.go:368 +0xbc
Filter through log files | View test history on testgrid
go run hack/e2e.go -v --test --test_args='--ginkgo.focus=Kubernetes\se2e\ssuite\s\[It\]\s\[sig\-apps\]\sStatefulSet\sBasic\sStatefulSet\sfunctionality\s\[StatefulSetBasic\]\sBurst\sscaling\sshould\srun\sto\scompletion\seven\swith\sunhealthy\spods\s\[Slow\]\s\[Conformance\]$'
test/e2e/framework/framework.go:241 k8s.io/kubernetes/test/e2e/framework.(*Framework).BeforeEach(0xc00113e1e0) test/e2e/framework/framework.go:241 +0x96ffrom junit_01.xml
[BeforeEach] [sig-apps] StatefulSet set up framework | framework.go:178 STEP: Creating a kubernetes client 11/26/22 05:31:47.666 Nov 26 05:31:47.666: INFO: >>> kubeConfig: /workspace/.kube/config STEP: Building a namespace api object, basename statefulset 11/26/22 05:31:47.668 Nov 26 05:31:47.707: INFO: Unexpected error while creating namespace: Post "https://35.227.179.144/api/v1/namespaces": dial tcp 35.227.179.144:443: connect: connection refused Nov 26 05:31:49.747: INFO: Unexpected error while creating namespace: Post "https://35.227.179.144/api/v1/namespaces": dial tcp 35.227.179.144:443: connect: connection refused Nov 26 05:31:51.747: INFO: Unexpected error while creating namespace: Post "https://35.227.179.144/api/v1/namespaces": dial tcp 35.227.179.144:443: connect: connection refused Nov 26 05:31:53.747: INFO: Unexpected error while creating namespace: Post "https://35.227.179.144/api/v1/namespaces": dial tcp 35.227.179.144:443: connect: connection refused Nov 26 05:31:55.747: INFO: Unexpected error while creating namespace: Post "https://35.227.179.144/api/v1/namespaces": dial tcp 35.227.179.144:443: connect: connection refused Nov 26 05:31:57.747: INFO: Unexpected error while creating namespace: Post "https://35.227.179.144/api/v1/namespaces": dial tcp 35.227.179.144:443: connect: connection refused Nov 26 05:31:59.747: INFO: Unexpected error while creating namespace: Post "https://35.227.179.144/api/v1/namespaces": dial tcp 35.227.179.144:443: connect: connection refused Nov 26 05:32:01.747: INFO: Unexpected error while creating namespace: Post "https://35.227.179.144/api/v1/namespaces": dial tcp 35.227.179.144:443: connect: connection refused Nov 26 05:32:03.747: INFO: Unexpected error while creating namespace: Post "https://35.227.179.144/api/v1/namespaces": dial tcp 35.227.179.144:443: connect: connection refused Nov 26 05:32:05.747: INFO: Unexpected error while creating namespace: Post "https://35.227.179.144/api/v1/namespaces": dial tcp 35.227.179.144:443: connect: connection refused Nov 26 05:32:07.747: INFO: Unexpected error while creating namespace: Post "https://35.227.179.144/api/v1/namespaces": dial tcp 35.227.179.144:443: connect: connection refused Nov 26 05:32:09.747: INFO: Unexpected error while creating namespace: Post "https://35.227.179.144/api/v1/namespaces": dial tcp 35.227.179.144:443: connect: connection refused Nov 26 05:32:11.747: INFO: Unexpected error while creating namespace: Post "https://35.227.179.144/api/v1/namespaces": dial tcp 35.227.179.144:443: connect: connection refused Nov 26 05:32:13.746: INFO: Unexpected error while creating namespace: Post "https://35.227.179.144/api/v1/namespaces": dial tcp 35.227.179.144:443: connect: connection refused Nov 26 05:32:15.747: INFO: Unexpected error while creating namespace: Post "https://35.227.179.144/api/v1/namespaces": dial tcp 35.227.179.144:443: connect: connection refused Nov 26 05:32:17.747: INFO: Unexpected error while creating namespace: Post "https://35.227.179.144/api/v1/namespaces": dial tcp 35.227.179.144:443: connect: connection refused Nov 26 05:32:17.786: INFO: Unexpected error while creating namespace: Post "https://35.227.179.144/api/v1/namespaces": dial tcp 35.227.179.144:443: connect: connection refused Nov 26 05:32:17.786: INFO: Unexpected error: <*errors.errorString | 0xc0000d1cd0>: { s: "timed out waiting for the condition", } Nov 26 05:32:17.786: FAIL: timed out waiting for the condition Full Stack Trace k8s.io/kubernetes/test/e2e/framework.(*Framework).BeforeEach(0xc00113e1e0) test/e2e/framework/framework.go:241 +0x96f [AfterEach] [sig-apps] StatefulSet test/e2e/framework/node/init/init.go:32 Nov 26 05:32:17.786: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready [DeferCleanup (Each)] [sig-apps] StatefulSet dump namespaces | framework.go:196 STEP: dump namespace information after failure 11/26/22 05:32:17.825 [DeferCleanup (Each)] [sig-apps] StatefulSet tear down framework | framework.go:193
Filter through log files | View test history on testgrid
go run hack/e2e.go -v --test --test_args='--ginkgo.focus=Kubernetes\se2e\ssuite\s\[It\]\s\[sig\-apps\]\sStatefulSet\sBasic\sStatefulSet\sfunctionality\s\[StatefulSetBasic\]\sScaling\sshould\shappen\sin\spredictable\sorder\sand\shalt\sif\sany\sstateful\spod\sis\sunhealthy\s\[Slow\]\s\[Conformance\]$'
test/e2e/framework/framework.go:241 k8s.io/kubernetes/test/e2e/framework.(*Framework).BeforeEach(0xc0007da1e0) test/e2e/framework/framework.go:241 +0x96ffrom junit_01.xml
[BeforeEach] [sig-apps] StatefulSet set up framework | framework.go:178 STEP: Creating a kubernetes client 11/26/22 05:32:17.883 Nov 26 05:32:17.884: INFO: >>> kubeConfig: /workspace/.kube/config STEP: Building a namespace api object, basename statefulset 11/26/22 05:32:17.885 Nov 26 05:32:17.925: INFO: Unexpected error while creating namespace: Post "https://35.227.179.144/api/v1/namespaces": dial tcp 35.227.179.144:443: connect: connection refused Nov 26 05:32:19.964: INFO: Unexpected error while creating namespace: Post "https://35.227.179.144/api/v1/namespaces": dial tcp 35.227.179.144:443: connect: connection refused Nov 26 05:32:21.964: INFO: Unexpected error while creating namespace: Post "https://35.227.179.144/api/v1/namespaces": dial tcp 35.227.179.144:443: connect: connection refused Nov 26 05:32:23.965: INFO: Unexpected error while creating namespace: Post "https://35.227.179.144/api/v1/namespaces": dial tcp 35.227.179.144:443: connect: connection refused Nov 26 05:32:25.964: INFO: Unexpected error while creating namespace: Post "https://35.227.179.144/api/v1/namespaces": dial tcp 35.227.179.144:443: connect: connection refused Nov 26 05:34:31.175: INFO: Unexpected error: <*fmt.wrapError | 0xc000280040>: { msg: "wait for service account \"default\" in namespace \"statefulset-3690\": timed out waiting for the condition", err: <*errors.errorString | 0xc00017da30>{ s: "timed out waiting for the condition", }, } Nov 26 05:34:31.175: FAIL: wait for service account "default" in namespace "statefulset-3690": timed out waiting for the condition Full Stack Trace k8s.io/kubernetes/test/e2e/framework.(*Framework).BeforeEach(0xc0007da1e0) test/e2e/framework/framework.go:241 +0x96f [AfterEach] [sig-apps] StatefulSet test/e2e/framework/node/init/init.go:32 Nov 26 05:34:31.175: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready [DeferCleanup (Each)] [sig-apps] StatefulSet dump namespaces | framework.go:196 STEP: dump namespace information after failure 11/26/22 05:34:31.273 STEP: Collecting events from namespace "statefulset-3690". 11/26/22 05:34:31.273 STEP: Found 0 events. 11/26/22 05:34:31.322 Nov 26 05:34:31.372: INFO: POD NODE PHASE GRACE CONDITIONS Nov 26 05:34:31.372: INFO: Nov 26 05:34:31.430: INFO: Logging node info for node bootstrap-e2e-master Nov 26 05:34:31.506: INFO: Node Info: &Node{ObjectMeta:{bootstrap-e2e-master 9a48eb11-c964-4bf8-82c0-1c32a9703981 12024 0 2022-11-26 05:05:18 +0000 UTC <nil> <nil> map[beta.kubernetes.io/arch:amd64 beta.kubernetes.io/instance-type:n1-standard-1 beta.kubernetes.io/os:linux cloud.google.com/metadata-proxy-ready:true failure-domain.beta.kubernetes.io/region:us-west1 failure-domain.beta.kubernetes.io/zone:us-west1-b kubernetes.io/arch:amd64 kubernetes.io/hostname:bootstrap-e2e-master kubernetes.io/os:linux node.kubernetes.io/instance-type:n1-standard-1 topology.kubernetes.io/region:us-west1 topology.kubernetes.io/zone:us-west1-b] map[node.alpha.kubernetes.io/ttl:0 volumes.kubernetes.io/controller-managed-attach-detach:true] [] [] [{kubelet Update v1 2022-11-26 05:05:18 +0000 UTC FieldsV1 {"f:metadata":{"f:annotations":{".":{},"f:volumes.kubernetes.io/controller-managed-attach-detach":{}},"f:labels":{".":{},"f:beta.kubernetes.io/arch":{},"f:beta.kubernetes.io/instance-type":{},"f:beta.kubernetes.io/os":{},"f:cloud.google.com/metadata-proxy-ready":{},"f:failure-domain.beta.kubernetes.io/region":{},"f:failure-domain.beta.kubernetes.io/zone":{},"f:kubernetes.io/arch":{},"f:kubernetes.io/hostname":{},"f:kubernetes.io/os":{},"f:node.kubernetes.io/instance-type":{},"f:topology.kubernetes.io/region":{},"f:topology.kubernetes.io/zone":{}}},"f:spec":{"f:providerID":{},"f:unschedulable":{}}} } {kube-controller-manager Update v1 2022-11-26 05:05:25 +0000 UTC FieldsV1 {"f:status":{"f:conditions":{"k:{\"type\":\"NetworkUnavailable\"}":{"f:lastHeartbeatTime":{},"f:lastTransitionTime":{},"f:message":{},"f:reason":{},"f:status":{}}}}} status} {kube-controller-manager Update v1 2022-11-26 05:05:39 +0000 UTC FieldsV1 {"f:metadata":{"f:annotations":{"f:node.alpha.kubernetes.io/ttl":{}}},"f:spec":{"f:podCIDR":{},"f:podCIDRs":{".":{},"v:\"10.64.1.0/24\"":{}},"f:taints":{}}} } {kubelet Update v1 2022-11-26 05:32:43 +0000 UTC FieldsV1 {"f:status":{"f:conditions":{"k:{\"type\":\"DiskPressure\"}":{"f:lastHeartbeatTime":{}},"k:{\"type\":\"MemoryPressure\"}":{"f:lastHeartbeatTime":{}},"k:{\"type\":\"PIDPressure\"}":{"f:lastHeartbeatTime":{}},"k:{\"type\":\"Ready\"}":{"f:lastHeartbeatTime":{},"f:lastTransitionTime":{},"f:message":{},"f:reason":{},"f:status":{}}},"f:images":{}}} status}]},Spec:NodeSpec{PodCIDR:10.64.1.0/24,DoNotUseExternalID:,ProviderID:gce://k8s-jkns-e2e-gce-ci-reboot/us-west1-b/bootstrap-e2e-master,Unschedulable:true,Taints:[]Taint{Taint{Key:node-role.kubernetes.io/master,Value:,Effect:NoSchedule,TimeAdded:<nil>,},Taint{Key:node.kubernetes.io/unschedulable,Value:,Effect:NoSchedule,TimeAdded:<nil>,},},ConfigSource:nil,PodCIDRs:[10.64.1.0/24],},Status:NodeStatus{Capacity:ResourceList{attachable-volumes-gce-pd: {{127 0} {<nil>} 127 DecimalSI},cpu: {{1 0} {<nil>} 1 DecimalSI},ephemeral-storage: {{16656896000 0} {<nil>} 16266500Ki BinarySI},hugepages-1Gi: {{0 0} {<nil>} 0 DecimalSI},hugepages-2Mi: {{0 0} {<nil>} 0 DecimalSI},memory: {{3858374656 0} {<nil>} BinarySI},pods: {{110 0} {<nil>} 110 DecimalSI},},Allocatable:ResourceList{attachable-volumes-gce-pd: {{127 0} {<nil>} 127 DecimalSI},cpu: {{1 0} {<nil>} 1 DecimalSI},ephemeral-storage: {{14991206376 0} {<nil>} 14991206376 DecimalSI},hugepages-1Gi: {{0 0} {<nil>} 0 DecimalSI},hugepages-2Mi: {{0 0} {<nil>} 0 DecimalSI},memory: {{3596230656 0} {<nil>} BinarySI},pods: {{110 0} {<nil>} 110 DecimalSI},},Phase:,Conditions:[]NodeCondition{NodeCondition{Type:NetworkUnavailable,Status:False,LastHeartbeatTime:2022-11-26 05:05:25 +0000 UTC,LastTransitionTime:2022-11-26 05:05:25 +0000 UTC,Reason:RouteCreated,Message:RouteController created a route,},NodeCondition{Type:MemoryPressure,Status:False,LastHeartbeatTime:2022-11-26 05:32:43 +0000 UTC,LastTransitionTime:2022-11-26 05:05:18 +0000 UTC,Reason:KubeletHasSufficientMemory,Message:kubelet has sufficient memory available,},NodeCondition{Type:DiskPressure,Status:False,LastHeartbeatTime:2022-11-26 05:32:43 +0000 UTC,LastTransitionTime:2022-11-26 05:05:18 +0000 UTC,Reason:KubeletHasNoDiskPressure,Message:kubelet has no disk pressure,},NodeCondition{Type:PIDPressure,Status:False,LastHeartbeatTime:2022-11-26 05:32:43 +0000 UTC,LastTransitionTime:2022-11-26 05:05:18 +0000 UTC,Reason:KubeletHasSufficientPID,Message:kubelet has sufficient PID available,},NodeCondition{Type:Ready,Status:True,LastHeartbeatTime:2022-11-26 05:32:43 +0000 UTC,LastTransitionTime:2022-11-26 05:05:38 +0000 UTC,Reason:KubeletReady,Message:kubelet is posting ready status. AppArmor enabled,},},Addresses:[]NodeAddress{NodeAddress{Type:InternalIP,Address:10.138.0.2,},NodeAddress{Type:ExternalIP,Address:35.227.179.144,},NodeAddress{Type:InternalDNS,Address:bootstrap-e2e-master.c.k8s-jkns-e2e-gce-ci-reboot.internal,},NodeAddress{Type:Hostname,Address:bootstrap-e2e-master.c.k8s-jkns-e2e-gce-ci-reboot.internal,},},DaemonEndpoints:NodeDaemonEndpoints{KubeletEndpoint:DaemonEndpoint{Port:10250,},},NodeInfo:NodeSystemInfo{MachineID:0fdb3cfe29f66637553465718381a2f8,SystemUUID:0fdb3cfe-29f6-6637-5534-65718381a2f8,BootID:fb804a29-7e61-434d-8044-6944159ce35f,KernelVersion:5.10.123+,OSImage:Container-Optimized OS from Google,ContainerRuntimeVersion:containerd://1.7.0-beta.0-149-gd06318622,KubeletVersion:v1.27.0-alpha.0.50+70617042976dc1,KubeProxyVersion:v1.27.0-alpha.0.50+70617042976dc1,OperatingSystem:linux,Architecture:amd64,},Images:[]ContainerImage{ContainerImage{Names:[registry.k8s.io/kube-apiserver-amd64:v1.27.0-alpha.0.50_70617042976dc1],SizeBytes:135160272,},ContainerImage{Names:[registry.k8s.io/kube-controller-manager-amd64:v1.27.0-alpha.0.50_70617042976dc1],SizeBytes:124990265,},ContainerImage{Names:[registry.k8s.io/etcd@sha256:dd75ec974b0a2a6f6bb47001ba09207976e625db898d1b16735528c009cb171c registry.k8s.io/etcd:3.5.6-0],SizeBytes:102542580,},ContainerImage{Names:[registry.k8s.io/kube-scheduler-amd64:v1.27.0-alpha.0.50_70617042976dc1],SizeBytes:57660216,},ContainerImage{Names:[gke.gcr.io/prometheus-to-sd@sha256:e739643c3939ba0b161425f45a1989eedfc4a3b166db9a7100863296b4c70510 gke.gcr.io/prometheus-to-sd:v0.11.1-gke.1],SizeBytes:48742566,},ContainerImage{Names:[gcr.io/k8s-ingress-image-push/ingress-gce-glbc-amd64@sha256:5db27383add6d9f4ebdf0286409ac31f7f5d273690204b341a4e37998917693b gcr.io/k8s-ingress-image-push/ingress-gce-glbc-amd64:v1.20.1],SizeBytes:36598135,},ContainerImage{Names:[registry.k8s.io/addon-manager/kube-addon-manager@sha256:49cc4e6e4a3745b427ce14b0141476ab339bb65c6bc05033019e046c8727dcb0 registry.k8s.io/addon-manager/kube-addon-manager:v9.1.6],SizeBytes:30464183,},ContainerImage{Names:[registry.k8s.io/kas-network-proxy/proxy-server@sha256:2c111f004bec24888d8cfa2a812a38fb8341350abac67dcd0ac64e709dfe389c registry.k8s.io/kas-network-proxy/proxy-server:v0.0.33],SizeBytes:22020129,},ContainerImage{Names:[registry.k8s.io/metadata-proxy@sha256:e914645f22e946bce5165737e1b244e0a296ad1f0f81a9531adc57af2780978a registry.k8s.io/metadata-proxy:v0.1.12],SizeBytes:5301657,},ContainerImage{Names:[registry.k8s.io/pause@sha256:9001185023633d17a2f98ff69b6ff2615b8ea02a825adffa40422f51dfdcde9d registry.k8s.io/pause:3.8],SizeBytes:311286,},},VolumesInUse:[],VolumesAttached:[]AttachedVolume{},Config:nil,},} Nov 26 05:34:31.507: INFO: Logging kubelet events for node bootstrap-e2e-master Nov 26 05:34:31.618: INFO: Logging pods the kubelet thinks is on node bootstrap-e2e-master Nov 26 05:34:32.155: INFO: etcd-server-events-bootstrap-e2e-master started at 2022-11-26 05:04:35 +0000 UTC (0+1 container statuses recorded) Nov 26 05:34:32.155: INFO: Container etcd-container ready: true, restart count 4 Nov 26 05:34:32.155: INFO: kube-controller-manager-bootstrap-e2e-master started at 2022-11-26 05:04:35 +0000 UTC (0+1 container statuses recorded) Nov 26 05:34:32.155: INFO: Container kube-controller-manager ready: false, restart count 7 Nov 26 05:34:32.155: INFO: kube-scheduler-bootstrap-e2e-master started at 2022-11-26 05:04:35 +0000 UTC (0+1 container statuses recorded) Nov 26 05:34:32.155: INFO: Container kube-scheduler ready: false, restart count 7 Nov 26 05:34:32.155: INFO: kube-addon-manager-bootstrap-e2e-master started at 2022-11-26 05:04:52 +0000 UTC (0+1 container statuses recorded) Nov 26 05:34:32.155: INFO: Container kube-addon-manager ready: true, restart count 1 Nov 26 05:34:32.155: INFO: metadata-proxy-v0.1-8ll2g started at 2022-11-26 05:05:51 +0000 UTC (0+2 container statuses recorded) Nov 26 05:34:32.156: INFO: Container metadata-proxy ready: true, restart count 0 Nov 26 05:34:32.156: INFO: Container prometheus-to-sd-exporter ready: true, restart count 0 Nov 26 05:34:32.156: INFO: etcd-server-bootstrap-e2e-master started at 2022-11-26 05:04:35 +0000 UTC (0+1 container statuses recorded) Nov 26 05:34:32.156: INFO: Container etcd-container ready: true, restart count 5 Nov 26 05:34:32.156: INFO: konnectivity-server-bootstrap-e2e-master started at 2022-11-26 05:04:35 +0000 UTC (0+1 container statuses recorded) Nov 26 05:34:32.156: INFO: Container konnectivity-server-container ready: true, restart count 3 Nov 26 05:34:32.156: INFO: kube-apiserver-bootstrap-e2e-master started at 2022-11-26 05:04:35 +0000 UTC (0+1 container statuses recorded) Nov 26 05:34:32.156: INFO: Container kube-apiserver ready: true, restart count 2 Nov 26 05:34:32.156: INFO: l7-lb-controller-bootstrap-e2e-master started at 2022-11-26 05:04:52 +0000 UTC (0+1 container statuses recorded) Nov 26 05:34:32.156: INFO: Container l7-lb-controller ready: false, restart count 8 Nov 26 05:34:33.489: INFO: Latency metrics for node bootstrap-e2e-master Nov 26 05:34:33.489: INFO: Logging node info for node bootstrap-e2e-minion-group-0975 Nov 26 05:34:33.538: INFO: Node Info: &Node{ObjectMeta:{bootstrap-e2e-minion-group-0975 6a0671d4-0e9d-4ac8-b359-65111262cfad 12145 0 2022-11-26 05:05:22 +0000 UTC <nil> <nil> map[beta.kubernetes.io/arch:amd64 beta.kubernetes.io/instance-type:n1-standard-2 beta.kubernetes.io/os:linux cloud.google.com/metadata-proxy-ready:true failure-domain.beta.kubernetes.io/region:us-west1 failure-domain.beta.kubernetes.io/zone:us-west1-b kubernetes.io/arch:amd64 kubernetes.io/hostname:bootstrap-e2e-minion-group-0975 kubernetes.io/os:linux node.kubernetes.io/instance-type:n1-standard-2 topology.hostpath.csi/node:bootstrap-e2e-minion-group-0975 topology.kubernetes.io/region:us-west1 topology.kubernetes.io/zone:us-west1-b] map[csi.volume.kubernetes.io/nodeid:{"csi-hostpath-multivolume-2344":"bootstrap-e2e-minion-group-0975","csi-hostpath-multivolume-593":"bootstrap-e2e-minion-group-0975"} node.alpha.kubernetes.io/ttl:0 volumes.kubernetes.io/controller-managed-attach-detach:true] [] [] [{kubelet Update v1 2022-11-26 05:05:22 +0000 UTC FieldsV1 {"f:metadata":{"f:annotations":{".":{},"f:volumes.kubernetes.io/controller-managed-attach-detach":{}},"f:labels":{".":{},"f:beta.kubernetes.io/arch":{},"f:beta.kubernetes.io/instance-type":{},"f:beta.kubernetes.io/os":{},"f:cloud.google.com/metadata-proxy-ready":{},"f:failure-domain.beta.kubernetes.io/region":{},"f:failure-domain.beta.kubernetes.io/zone":{},"f:kubernetes.io/arch":{},"f:kubernetes.io/hostname":{},"f:kubernetes.io/os":{},"f:node.kubernetes.io/instance-type":{},"f:topology.kubernetes.io/region":{},"f:topology.kubernetes.io/zone":{}}},"f:spec":{"f:providerID":{}}} } {kube-controller-manager Update v1 2022-11-26 05:05:23 +0000 UTC FieldsV1 {"f:metadata":{"f:annotations":{"f:node.alpha.kubernetes.io/ttl":{}}},"f:spec":{"f:podCIDR":{},"f:podCIDRs":{".":{},"v:\"10.64.2.0/24\"":{}}}} } {kube-controller-manager Update v1 2022-11-26 05:13:31 +0000 UTC FieldsV1 {"f:status":{"f:conditions":{"k:{\"type\":\"NetworkUnavailable\"}":{"f:lastHeartbeatTime":{},"f:lastTransitionTime":{},"f:message":{},"f:reason":{},"f:status":{}}}}} status} {node-problem-detector Update v1 2022-11-26 05:30:31 +0000 UTC FieldsV1 {"f:status":{"f:conditions":{"k:{\"type\":\"CorruptDockerOverlay2\"}":{".":{},"f:lastHeartbeatTime":{},"f:lastTransitionTime":{},"f:message":{},"f:reason":{},"f:status":{},"f:type":{}},"k:{\"type\":\"FrequentContainerdRestart\"}":{".":{},"f:lastHeartbeatTime":{},"f:lastTransitionTime":{},"f:message":{},"f:reason":{},"f:status":{},"f:type":{}},"k:{\"type\":\"FrequentDockerRestart\"}":{".":{},"f:lastHeartbeatTime":{},"f:lastTransitionTime":{},"f:message":{},"f:reason":{},"f:status":{},"f:type":{}},"k:{\"type\":\"FrequentKubeletRestart\"}":{".":{},"f:lastHeartbeatTime":{},"f:lastTransitionTime":{},"f:message":{},"f:reason":{},"f:status":{},"f:type":{}},"k:{\"type\":\"FrequentUnregisterNetDevice\"}":{".":{},"f:lastHeartbeatTime":{},"f:lastTransitionTime":{},"f:message":{},"f:reason":{},"f:status":{},"f:type":{}},"k:{\"type\":\"KernelDeadlock\"}":{".":{},"f:lastHeartbeatTime":{},"f:lastTransitionTime":{},"f:message":{},"f:reason":{},"f:status":{},"f:type":{}},"k:{\"type\":\"ReadonlyFilesystem\"}":{".":{},"f:lastHeartbeatTime":{},"f:lastTransitionTime":{},"f:message":{},"f:reason":{},"f:status":{},"f:type":{}}}}} status} {kubelet Update v1 2022-11-26 05:33:41 +0000 UTC FieldsV1 {"f:metadata":{"f:annotations":{"f:csi.volume.kubernetes.io/nodeid":{}},"f:labels":{"f:topology.hostpath.csi/node":{}}},"f:status":{"f:conditions":{"k:{\"type\":\"DiskPressure\"}":{"f:lastHeartbeatTime":{}},"k:{\"type\":\"MemoryPressure\"}":{"f:lastHeartbeatTime":{}},"k:{\"type\":\"PIDPressure\"}":{"f:lastHeartbeatTime":{}},"k:{\"type\":\"Ready\"}":{"f:lastHeartbeatTime":{},"f:lastTransitionTime":{},"f:message":{},"f:reason":{},"f:status":{}}},"f:images":{}}} status}]},Spec:NodeSpec{PodCIDR:10.64.2.0/24,DoNotUseExternalID:,ProviderID:gce://k8s-jkns-e2e-gce-ci-reboot/us-west1-b/bootstrap-e2e-minion-group-0975,Unschedulable:false,Taints:[]Taint{},ConfigSource:nil,PodCIDRs:[10.64.2.0/24],},Status:NodeStatus{Capacity:ResourceList{attachable-volumes-gce-pd: {{127 0} {<nil>} 127 DecimalSI},cpu: {{2 0} {<nil>} 2 DecimalSI},ephemeral-storage: {{101203873792 0} {<nil>} 98831908Ki BinarySI},hugepages-1Gi: {{0 0} {<nil>} 0 DecimalSI},hugepages-2Mi: {{0 0} {<nil>} 0 DecimalSI},memory: {{7815430144 0} {<nil>} BinarySI},pods: {{110 0} {<nil>} 110 DecimalSI},},Allocatable:ResourceList{attachable-volumes-gce-pd: {{127 0} {<nil>} 127 DecimalSI},cpu: {{2 0} {<nil>} 2 DecimalSI},ephemeral-storage: {{91083486262 0} {<nil>} 91083486262 DecimalSI},hugepages-1Gi: {{0 0} {<nil>} 0 DecimalSI},hugepages-2Mi: {{0 0} {<nil>} 0 DecimalSI},memory: {{7553286144 0} {<nil>} BinarySI},pods: {{110 0} {<nil>} 110 DecimalSI},},Phase:,Conditions:[]NodeCondition{NodeCondition{Type:CorruptDockerOverlay2,Status:False,LastHeartbeatTime:2022-11-26 05:30:31 +0000 UTC,LastTransitionTime:2022-11-26 05:05:27 +0000 UTC,Reason:NoCorruptDockerOverlay2,Message:docker overlay2 is functioning properly,},NodeCondition{Type:FrequentUnregisterNetDevice,Status:False,LastHeartbeatTime:2022-11-26 05:30:31 +0000 UTC,LastTransitionTime:2022-11-26 05:05:27 +0000 UTC,Reason:NoFrequentUnregisterNetDevice,Message:node is functioning properly,},NodeCondition{Type:FrequentKubeletRestart,Status:False,LastHeartbeatTime:2022-11-26 05:30:31 +0000 UTC,LastTransitionTime:2022-11-26 05:05:27 +0000 UTC,Reason:NoFrequentKubeletRestart,Message:kubelet is functioning properly,},NodeCondition{Type:FrequentDockerRestart,Status:False,LastHeartbeatTime:2022-11-26 05:30:31 +0000 UTC,LastTransitionTime:2022-11-26 05:05:27 +0000 UTC,Reason:NoFrequentDockerRestart,Message:docker is functioning properly,},NodeCondition{Type:FrequentContainerdRestart,Status:False,LastHeartbeatTime:2022-11-26 05:30:31 +0000 UTC,LastTransitionTime:2022-11-26 05:05:27 +0000 UTC,Reason:NoFrequentContainerdRestart,Message:containerd is functioning properly,},NodeCondition{Type:KernelDeadlock,Status:False,LastHeartbeatTime:2022-11-26 05:30:31 +0000 UTC,LastTransitionTime:2022-11-26 05:05:27 +0000 UTC,Reason:KernelHasNoDeadlock,Message:kernel has no deadlock,},NodeCondition{Type:ReadonlyFilesystem,Status:False,LastHeartbeatTime:2022-11-26 05:30:31 +0000 UTC,LastTransitionTime:2022-11-26 05:05:27 +0000 UTC,Reason:FilesystemIsNotReadOnly,Message:Filesystem is not read-only,},NodeCondition{Type:NetworkUnavailable,Status:False,LastHeartbeatTime:2022-11-26 05:05:39 +0000 UTC,LastTransitionTime:2022-11-26 05:05:39 +0000 UTC,Reason:RouteCreated,Message:RouteController created a route,},NodeCondition{Type:MemoryPressure,Status:False,LastHeartbeatTime:2022-11-26 05:29:28 +0000 UTC,LastTransitionTime:2022-11-26 05:05:22 +0000 UTC,Reason:KubeletHasSufficientMemory,Message:kubelet has sufficient memory available,},NodeCondition{Type:DiskPressure,Status:False,LastHeartbeatTime:2022-11-26 05:29:28 +0000 UTC,LastTransitionTime:2022-11-26 05:05:22 +0000 UTC,Reason:KubeletHasNoDiskPressure,Message:kubelet has no disk pressure,},NodeCondition{Type:PIDPressure,Status:False,LastHeartbeatTime:2022-11-26 05:29:28 +0000 UTC,LastTransitionTime:2022-11-26 05:05:22 +0000 UTC,Reason:KubeletHasSufficientPID,Message:kubelet has sufficient PID available,},NodeCondition{Type:Ready,Status:True,LastHeartbeatTime:2022-11-26 05:29:28 +0000 UTC,LastTransitionTime:2022-11-26 05:05:23 +0000 UTC,Reason:KubeletReady,Message:kubelet is posting ready status. AppArmor enabled,},},Addresses:[]NodeAddress{NodeAddress{Type:InternalIP,Address:10.138.0.5,},NodeAddress{Type:ExternalIP,Address:35.247.104.236,},NodeAddress{Type:InternalDNS,Address:bootstrap-e2e-minion-group-0975.c.k8s-jkns-e2e-gce-ci-reboot.internal,},NodeAddress{Type:Hostname,Address:bootstrap-e2e-minion-group-0975.c.k8s-jkns-e2e-gce-ci-reboot.internal,},},DaemonEndpoints:NodeDaemonEndpoints{KubeletEndpoint:DaemonEndpoint{Port:10250,},},NodeInfo:NodeSystemInfo{MachineID:b38385f7976a474936da82f5f565760a,SystemUUID:b38385f7-976a-4749-36da-82f5f565760a,BootID:0aa74273-9b8d-48ac-bad0-0a1805f010ea,KernelVersion:5.10.123+,OSImage:Container-Optimized OS from Google,ContainerRuntimeVersion:containerd://1.7.0-beta.0-149-gd06318622,KubeletVersion:v1.27.0-alpha.0.50+70617042976dc1,KubeProxyVersion:v1.27.0-alpha.0.50+70617042976dc1,OperatingSystem:linux,Architecture:amd64,},Images:[]ContainerImage{ContainerImage{Names:[registry.k8s.io/e2e-test-images/jessie-dnsutils@sha256:24aaf2626d6b27864c29de2097e8bbb840b3a414271bf7c8995e431e47d8408e registry.k8s.io/e2e-test-images/jessie-dnsutils:1.7],SizeBytes:112030336,},ContainerImage{Names:[registry.k8s.io/e2e-test-images/volume/nfs@sha256:3bda73f2428522b0e342af80a0b9679e8594c2126f2b3cca39ed787589741b9e registry.k8s.io/e2e-test-images/volume/nfs:1.3],SizeBytes:95836203,},ContainerImage{Names:[registry.k8s.io/sig-storage/nfs-provisioner@sha256:e943bb77c7df05ebdc8c7888b2db289b13bf9f012d6a3a5a74f14d4d5743d439 registry.k8s.io/sig-storage/nfs-provisioner:v3.0.1],SizeBytes:90632047,},ContainerImage{Names:[registry.k8s.io/kube-proxy-amd64:v1.27.0-alpha.0.50_70617042976dc1],SizeBytes:67201736,},ContainerImage{Names:[registry.k8s.io/e2e-test-images/agnhost@sha256:16bbf38c463a4223d8cfe4da12bc61010b082a79b4bb003e2d3ba3ece5dd5f9e registry.k8s.io/e2e-test-images/agnhost:2.43],SizeBytes:51706353,},ContainerImage{Names:[gke.gcr.io/prometheus-to-sd@sha256:e739643c3939ba0b161425f45a1989eedfc4a3b166db9a7100863296b4c70510 gke.gcr.io/prometheus-to-sd:v0.11.1-gke.1],SizeBytes:48742566,},ContainerImage{Names:[registry.k8s.io/e2e-test-images/httpd@sha256:148b022f5c5da426fc2f3c14b5c0867e58ef05961510c84749ac1fddcb0fef22 registry.k8s.io/e2e-test-images/httpd:2.4.38-4],SizeBytes:40764257,},ContainerImage{Names:[registry.k8s.io/metrics-server/metrics-server@sha256:6385aec64bb97040a5e692947107b81e178555c7a5b71caa90d733e4130efc10 registry.k8s.io/metrics-server/metrics-server:v0.5.2],SizeBytes:26023008,},ContainerImage{Names:[registry.k8s.io/sig-storage/csi-provisioner@sha256:ee3b525d5b89db99da3b8eb521d9cd90cb6e9ef0fbb651e98bb37be78d36b5b8 registry.k8s.io/sig-storage/csi-provisioner:v3.3.0],SizeBytes:25491225,},ContainerImage{Names:[registry.k8s.io/sig-storage/csi-resizer@sha256:425d8f1b769398127767b06ed97ce62578a3179bcb99809ce93a1649e025ffe7 registry.k8s.io/sig-storage/csi-resizer:v1.6.0],SizeBytes:24148884,},ContainerImage{Names:[registry.k8s.io/sig-storage/csi-snapshotter@sha256:291334908ddf71a4661fd7f6d9d97274de8a5378a2b6fdfeb2ce73414a34f82f registry.k8s.io/sig-storage/csi-snapshotter:v6.1.0],SizeBytes:23881995,},ContainerImage{Names:[registry.k8s.io/sig-storage/csi-attacher@sha256:9a685020911e2725ad019dbce6e4a5ab93d51e3d4557f115e64343345e05781b registry.k8s.io/sig-storage/csi-attacher:v4.0.0],SizeBytes:23847201,},ContainerImage{Names:[registry.k8s.io/sig-storage/hostpathplugin@sha256:92257881c1d6493cf18299a24af42330f891166560047902b8d431fb66b01af5 registry.k8s.io/sig-storage/hostpathplugin:v1.9.0],SizeBytes:18758628,},ContainerImage{Names:[registry.k8s.io/autoscaling/addon-resizer@sha256:43f129b81d28f0fdd54de6d8e7eacd5728030782e03db16087fc241ad747d3d6 registry.k8s.io/autoscaling/addon-resizer:1.8.14],SizeBytes:10153852,},ContainerImage{Names:[registry.k8s.io/sig-storage/csi-node-driver-registrar@sha256:0103eee7c35e3e0b5cd8cdca9850dc71c793cdeb6669d8be7a89440da2d06ae4 registry.k8s.io/sig-storage/csi-node-driver-registrar:v2.5.1],SizeBytes:9133109,},ContainerImage{Names:[registry.k8s.io/sig-storage/livenessprobe@sha256:933940f13b3ea0abc62e656c1aa5c5b47c04b15d71250413a6b821bd0c58b94e registry.k8s.io/sig-storage/livenessprobe:v2.7.0],SizeBytes:8688564,},ContainerImage{Names:[registry.k8s.io/kas-network-proxy/proxy-agent@sha256:48f2a4ec3e10553a81b8dd1c6fa5fe4bcc9617f78e71c1ca89c6921335e2d7da registry.k8s.io/kas-network-proxy/proxy-agent:v0.0.33],SizeBytes:8512162,},ContainerImage{Names:[registry.k8s.io/metadata-proxy@sha256:e914645f22e946bce5165737e1b244e0a296ad1f0f81a9531adc57af2780978a registry.k8s.io/metadata-proxy:v0.1.12],SizeBytes:5301657,},ContainerImage{Names:[registry.k8s.io/e2e-test-images/busybox@sha256:c318242786b139d18676b1c09a0ad7f15fc17f8f16a5b2e625cd0dc8c9703daf registry.k8s.io/e2e-test-images/busybox:1.29-2],SizeBytes:732424,},ContainerImage{Names:[registry.k8s.io/e2e-test-images/busybox@sha256:2e0f836850e09b8b7cc937681d6194537a09fbd5f6b9e08f4d646a85128e8937 registry.k8s.io/e2e-test-images/busybox:1.29-4],SizeBytes:731990,},ContainerImage{Names:[registry.k8s.io/pause@sha256:9001185023633d17a2f98ff69b6ff2615b8ea02a825adffa40422f51dfdcde9d registry.k8s.io/pause:3.8],SizeBytes:311286,},},VolumesInUse:[],VolumesAttached:[]AttachedVolume{},Config:nil,},} Nov 26 05:34:33.539: INFO: Logging kubelet events for node bootstrap-e2e-minion-group-0975 Nov 26 05:34:33.603: INFO: Logging pods the kubelet thinks is on node bootstrap-e2e-minion-group-0975 Nov 26 05:34:33.855: INFO: kube-proxy-bootstrap-e2e-minion-group-0975 started at 2022-11-26 05:05:23 +0000 UTC (0+1 container statuses recorded) Nov 26 05:34:33.855: INFO: Container kube-proxy ready: false, restart count 8 Nov 26 05:34:33.855: INFO: hostexec-bootstrap-e2e-minion-group-0975-d2gqf started at 2022-11-26 05:31:42 +0000 UTC (0+1 container statuses recorded) Nov 26 05:34:33.855: INFO: Container agnhost-container ready: true, restart count 2 Nov 26 05:34:33.855: INFO: netserver-0 started at 2022-11-26 05:11:29 +0000 UTC (0+1 container statuses recorded) Nov 26 05:34:33.855: INFO: Container webserver ready: false, restart count 6 Nov 26 05:34:33.855: INFO: metrics-server-v0.5.2-867b8754b9-27xcg started at 2022-11-26 05:05:51 +0000 UTC (0+2 container statuses recorded) Nov 26 05:34:33.855: INFO: Container metrics-server ready: false, restart count 8 Nov 26 05:34:33.855: INFO: Container metrics-server-nanny ready: false, restart count 9 Nov 26 05:34:33.855: INFO: net-tiers-svc-sj7rf started at 2022-11-26 05:17:11 +0000 UTC (0+1 container statuses recorded) Nov 26 05:34:33.855: INFO: Container netexec ready: true, restart count 5 Nov 26 05:34:33.855: INFO: reallocate-nodeport-test-njbll started at 2022-11-26 05:23:12 +0000 UTC (0+1 container statuses recorded) Nov 26 05:34:33.855: INFO: Container netexec ready: true, restart count 4 Nov 26 05:34:33.855: INFO: lb-internal-r5l4p started at 2022-11-26 05:16:06 +0000 UTC (0+1 container statuses recorded) Nov 26 05:34:33.855: INFO: Container netexec ready: true, restart count 1 Nov 26 05:34:33.855: INFO: csi-hostpathplugin-0 started at 2022-11-26 05:31:42 +0000 UTC (0+7 container statuses recorded) Nov 26 05:34:33.855: INFO: Container csi-attacher ready: true, restart count 1 Nov 26 05:34:33.855: INFO: Container csi-provisioner ready: true, restart count 1 Nov 26 05:34:33.855: INFO: Container csi-resizer ready: true, restart count 1 Nov 26 05:34:33.855: INFO: Container csi-snapshotter ready: true, restart count 1 Nov 26 05:34:33.855: INFO: Container hostpath ready: true, restart count 1 Nov 26 05:34:33.855: INFO: Container liveness-probe ready: true, restart count 1 Nov 26 05:34:33.855: INFO: Container node-driver-registrar ready: true, restart count 1 Nov 26 05:34:33.855: INFO: addon-reconcile-test-lhstw started at 2022-11-26 05:31:42 +0000 UTC (0+1 container statuses recorded) Nov 26 05:34:33.855: INFO: Container addon-reconcile-test ready: true, restart count 1 Nov 26 05:34:33.855: INFO: hostexec-bootstrap-e2e-minion-group-0975-xhbnp started at 2022-11-26 05:31:42 +0000 UTC (0+1 container statuses recorded) Nov 26 05:34:33.855: INFO: Container agnhost-container ready: true, restart count 0 Nov 26 05:34:33.855: INFO: netserver-0 started at 2022-11-26 05:09:51 +0000 UTC (0+1 container statuses recorded) Nov 26 05:34:33.855: INFO: Container webserver ready: true, restart count 7 Nov 26 05:34:33.855: INFO: csi-hostpathplugin-0 started at 2022-11-26 05:09:53 +0000 UTC (0+7 container statuses recorded) Nov 26 05:34:33.855: INFO: Container csi-attacher ready: true, restart count 4 Nov 26 05:34:33.855: INFO: Container csi-provisioner ready: true, restart count 4 Nov 26 05:34:33.855: INFO: Container csi-resizer ready: true, restart count 4 Nov 26 05:34:33.855: INFO: Container csi-snapshotter ready: true, restart count 4 Nov 26 05:34:33.855: INFO: Container hostpath ready: true, restart count 4 Nov 26 05:34:33.855: INFO: Container liveness-probe ready: true, restart count 4 Nov 26 05:34:33.855: INFO: Container node-driver-registrar ready: true, restart count 4 Nov 26 05:34:33.855: INFO: external-local-update-5xshb started at 2022-11-26 05:11:23 +0000 UTC (0+1 container statuses recorded) Nov 26 05:34:33.855: INFO: Container netexec ready: true, restart count 8 Nov 26 05:34:33.855: INFO: host-test-container-pod started at 2022-11-26 05:12:10 +0000 UTC (0+1 container statuses recorded) Nov 26 05:34:33.855: INFO: Container agnhost-container ready: true, restart count 6 Nov 26 05:34:33.855: INFO: forbid-27824003-x9rls started at 2022-11-26 05:23:10 +0000 UTC (0+1 container statuses recorded) Nov 26 05:34:33.855: INFO: Container c ready: false, restart count 0 Nov 26 05:34:33.855: INFO: metadata-proxy-v0.1-mznrc started at 2022-11-26 05:05:24 +0000 UTC (0+2 container statuses recorded) Nov 26 05:34:33.855: INFO: Container metadata-proxy ready: true, restart count 0 Nov 26 05:34:33.855: INFO: Container prometheus-to-sd-exporter ready: true, restart count 0 Nov 26 05:34:33.855: INFO: konnectivity-agent-wfc29 started at 2022-11-26 05:05:39 +0000 UTC (0+1 container statuses recorded) Nov 26 05:34:33.855: INFO: Container konnectivity-agent ready: true, restart count 9 Nov 26 05:34:33.855: INFO: csi-mockplugin-0 started at 2022-11-26 05:23:18 +0000 UTC (0+4 container statuses recorded) Nov 26 05:34:33.855: INFO: Container busybox ready: false, restart count 5 Nov 26 05:34:33.855: INFO: Container csi-provisioner ready: false, restart count 5 Nov 26 05:34:33.855: INFO: Container driver-registrar ready: false, restart count 6 Nov 26 05:34:33.855: INFO: Container mock ready: false, restart count 6 Nov 26 05:34:33.855: INFO: netserver-0 started at 2022-11-26 05:11:06 +0000 UTC (0+1 container statuses recorded) Nov 26 05:34:33.855: INFO: Container webserver ready: false, restart count 7 Nov 26 05:34:33.855: INFO: csi-hostpathplugin-0 started at 2022-11-26 05:12:10 +0000 UTC (0+7 container statuses recorded) Nov 26 05:34:33.855: INFO: Container csi-attacher ready: false, restart count 9 Nov 26 05:34:33.855: INFO: Container csi-provisioner ready: false, restart count 9 Nov 26 05:34:33.855: INFO: Container csi-resizer ready: false, restart count 9 Nov 26 05:34:33.855: INFO: Container csi-snapshotter ready: false, restart count 9 Nov 26 05:34:33.855: INFO: Container hostpath ready: false, restart count 9 Nov 26 05:34:33.855: INFO: Container liveness-probe ready: false, restart count 9 Nov 26 05:34:33.855: INFO: Container node-driver-registrar ready: false, restart count 9 Nov 26 05:34:33.855: INFO: pod-secrets-ae6e2110-abaa-4433-ab33-c417114f813f started at 2022-11-26 05:10:35 +0000 UTC (0+1 container statuses recorded) Nov 26 05:34:33.855: INFO: Container creates-volume-test ready: false, restart count 0 Nov 26 05:34:33.855: INFO: csi-hostpathplugin-0 started at 2022-11-26 05:13:04 +0000 UTC (0+7 container statuses recorded) Nov 26 05:34:33.855: INFO: Container csi-attacher ready: true, restart count 6 Nov 26 05:34:33.855: INFO: Container csi-provisioner ready: true, restart count 6 Nov 26 05:34:33.855: INFO: Container csi-resizer ready: true, restart count 6 Nov 26 05:34:33.855: INFO: Container csi-snapshotter ready: true, restart count 6 Nov 26 05:34:33.855: INFO: Container hostpath ready: true, restart count 6 Nov 26 05:34:33.855: INFO: Container liveness-probe ready: true, restart count 6 Nov 26 05:34:33.855: INFO: Container node-driver-registrar ready: true, restart count 6 Nov 26 05:34:35.290: INFO: Latency metrics for node bootstrap-e2e-minion-group-0975 Nov 26 05:34:35.290: INFO: Logging node info for node bootstrap-e2e-minion-group-rwsn Nov 26 05:34:35.341: INFO: Node Info: &Node{ObjectMeta:{bootstrap-e2e-minion-group-rwsn adedf225-b462-4305-a60c-9042d31ca1d2 12192 0 2022-11-26 05:05:15 +0000 UTC <nil> <nil> map[beta.kubernetes.io/arch:amd64 beta.kubernetes.io/instance-type:n1-standard-2 beta.kubernetes.io/os:linux cloud.google.com/metadata-proxy-ready:true failure-domain.beta.kubernetes.io/region:us-west1 failure-domain.beta.kubernetes.io/zone:us-west1-b kubernetes.io/arch:amd64 kubernetes.io/hostname:bootstrap-e2e-minion-group-rwsn kubernetes.io/os:linux node.kubernetes.io/instance-type:n1-standard-2 topology.hostpath.csi/node:bootstrap-e2e-minion-group-rwsn topology.kubernetes.io/region:us-west1 topology.kubernetes.io/zone:us-west1-b] map[csi.volume.kubernetes.io/nodeid:{"csi-hostpath-multivolume-3900":"bootstrap-e2e-minion-group-rwsn"} node.alpha.kubernetes.io/ttl:0 volumes.kubernetes.io/controller-managed-attach-detach:true] [] [] [{kubelet Update v1 2022-11-26 05:05:15 +0000 UTC FieldsV1 {"f:metadata":{"f:annotations":{".":{},"f:volumes.kubernetes.io/controller-managed-attach-detach":{}},"f:labels":{".":{},"f:beta.kubernetes.io/arch":{},"f:beta.kubernetes.io/instance-type":{},"f:beta.kubernetes.io/os":{},"f:cloud.google.com/metadata-proxy-ready":{},"f:failure-domain.beta.kubernetes.io/region":{},"f:failure-domain.beta.kubernetes.io/zone":{},"f:kubernetes.io/arch":{},"f:kubernetes.io/hostname":{},"f:kubernetes.io/os":{},"f:node.kubernetes.io/instance-type":{},"f:topology.kubernetes.io/region":{},"f:topology.kubernetes.io/zone":{}}},"f:spec":{"f:providerID":{}}} } {kube-controller-manager Update v1 2022-11-26 05:05:17 +0000 UTC FieldsV1 {"f:metadata":{"f:annotations":{"f:node.alpha.kubernetes.io/ttl":{}}},"f:spec":{"f:podCIDR":{},"f:podCIDRs":{".":{},"v:\"10.64.0.0/24\"":{}}}} } {kube-controller-manager Update v1 2022-11-26 05:29:28 +0000 UTC FieldsV1 {"f:status":{"f:conditions":{"k:{\"type\":\"NetworkUnavailable\"}":{"f:lastHeartbeatTime":{},"f:lastTransitionTime":{},"f:message":{},"f:reason":{},"f:status":{}}},"f:volumesAttached":{}}} status} {node-problem-detector Update v1 2022-11-26 05:30:22 +0000 UTC FieldsV1 {"f:status":{"f:conditions":{"k:{\"type\":\"CorruptDockerOverlay2\"}":{".":{},"f:lastHeartbeatTime":{},"f:lastTransitionTime":{},"f:message":{},"f:reason":{},"f:status":{},"f:type":{}},"k:{\"type\":\"FrequentContainerdRestart\"}":{".":{},"f:lastHeartbeatTime":{},"f:lastTransitionTime":{},"f:message":{},"f:reason":{},"f:status":{},"f:type":{}},"k:{\"type\":\"FrequentDockerRestart\"}":{".":{},"f:lastHeartbeatTime":{},"f:lastTransitionTime":{},"f:message":{},"f:reason":{},"f:status":{},"f:type":{}},"k:{\"type\":\"FrequentKubeletRestart\"}":{".":{},"f:lastHeartbeatTime":{},"f:lastTransitionTime":{},"f:message":{},"f:reason":{},"f:status":{},"f:type":{}},"k:{\"type\":\"FrequentUnregisterNetDevice\"}":{".":{},"f:lastHeartbeatTime":{},"f:lastTransitionTime":{},"f:message":{},"f:reason":{},"f:status":{},"f:type":{}},"k:{\"type\":\"KernelDeadlock\"}":{".":{},"f:lastHeartbeatTime":{},"f:lastTransitionTime":{},"f:message":{},"f:reason":{},"f:status":{},"f:type":{}},"k:{\"type\":\"ReadonlyFilesystem\"}":{".":{},"f:lastHeartbeatTime":{},"f:lastTransitionTime":{},"f:message":{},"f:reason":{},"f:status":{},"f:type":{}}}}} status} {kubelet Update v1 2022-11-26 05:34:03 +0000 UTC FieldsV1 {"f:metadata":{"f:annotations":{"f:csi.volume.kubernetes.io/nodeid":{}},"f:labels":{"f:topology.hostpath.csi/node":{}}},"f:status":{"f:conditions":{"k:{\"type\":\"DiskPressure\"}":{"f:lastHeartbeatTime":{}},"k:{\"type\":\"MemoryPressure\"}":{"f:lastHeartbeatTime":{}},"k:{\"type\":\"PIDPressure\"}":{"f:lastHeartbeatTime":{}},"k:{\"type\":\"Ready\"}":{"f:lastHeartbeatTime":{},"f:lastTransitionTime":{},"f:message":{},"f:reason":{},"f:status":{}}},"f:images":{},"f:volumesInUse":{}}} status}]},Spec:NodeSpec{PodCIDR:10.64.0.0/24,DoNotUseExternalID:,ProviderID:gce://k8s-jkns-e2e-gce-ci-reboot/us-west1-b/bootstrap-e2e-minion-group-rwsn,Unschedulable:false,Taints:[]Taint{},ConfigSource:nil,PodCIDRs:[10.64.0.0/24],},Status:NodeStatus{Capacity:ResourceList{attachable-volumes-gce-pd: {{127 0} {<nil>} 127 DecimalSI},cpu: {{2 0} {<nil>} 2 DecimalSI},ephemeral-storage: {{101203873792 0} {<nil>} 98831908Ki BinarySI},hugepages-1Gi: {{0 0} {<nil>} 0 DecimalSI},hugepages-2Mi: {{0 0} {<nil>} 0 DecimalSI},memory: {{7815438336 0} {<nil>} BinarySI},pods: {{110 0} {<nil>} 110 DecimalSI},},Allocatable:ResourceList{attachable-volumes-gce-pd: {{127 0} {<nil>} 127 DecimalSI},cpu: {{2 0} {<nil>} 2 DecimalSI},ephemeral-storage: {{91083486262 0} {<nil>} 91083486262 DecimalSI},hugepages-1Gi: {{0 0} {<nil>} 0 DecimalSI},hugepages-2Mi: {{0 0} {<nil>} 0 DecimalSI},memory: {{7553294336 0} {<nil>} BinarySI},pods: {{110 0} {<nil>} 110 DecimalSI},},Phase:,Conditions:[]NodeCondition{NodeCondition{Type:FrequentContainerdRestart,Status:False,LastHeartbeatTime:2022-11-26 05:30:22 +0000 UTC,LastTransitionTime:2022-11-26 05:05:18 +0000 UTC,Reason:NoFrequentContainerdRestart,Message:containerd is functioning properly,},NodeCondition{Type:KernelDeadlock,Status:False,LastHeartbeatTime:2022-11-26 05:30:22 +0000 UTC,LastTransitionTime:2022-11-26 05:05:18 +0000 UTC,Reason:KernelHasNoDeadlock,Message:kernel has no deadlock,},NodeCondition{Type:ReadonlyFilesystem,Status:False,LastHeartbeatTime:2022-11-26 05:30:22 +0000 UTC,LastTransitionTime:2022-11-26 05:05:18 +0000 UTC,Reason:FilesystemIsNotReadOnly,Message:Filesystem is not read-only,},NodeCondition{Type:CorruptDockerOverlay2,Status:False,LastHeartbeatTime:2022-11-26 05:30:22 +0000 UTC,LastTransitionTime:2022-11-26 05:05:18 +0000 UTC,Reason:NoCorruptDockerOverlay2,Message:docker overlay2 is functioning properly,},NodeCondition{Type:FrequentUnregisterNetDevice,Status:False,LastHeartbeatTime:2022-11-26 05:30:22 +0000 UTC,LastTransitionTime:2022-11-26 05:05:18 +0000 UTC,Reason:NoFrequentUnregisterNetDevice,Message:node is functioning properly,},NodeCondition{Type:FrequentKubeletRestart,Status:False,LastHeartbeatTime:2022-11-26 05:30:22 +0000 UTC,LastTransitionTime:2022-11-26 05:05:18 +0000 UTC,Reason:NoFrequentKubeletRestart,Message:kubelet is functioning properly,},NodeCondition{Type:FrequentDockerRestart,Status:False,LastHeartbeatTime:2022-11-26 05:30:22 +0000 UTC,LastTransitionTime:2022-11-26 05:05:18 +0000 UTC,Reason:NoFrequentDockerRestart,Message:docker is functioning properly,},NodeCondition{Type:NetworkUnavailable,Status:False,LastHeartbeatTime:2022-11-26 05:05:25 +0000 UTC,LastTransitionTime:2022-11-26 05:05:25 +0000 UTC,Reason:RouteCreated,Message:RouteController created a route,},NodeCondition{Type:MemoryPressure,Status:False,LastHeartbeatTime:2022-11-26 05:32:57 +0000 UTC,LastTransitionTime:2022-11-26 05:05:15 +0000 UTC,Reason:KubeletHasSufficientMemory,Message:kubelet has sufficient memory available,},NodeCondition{Type:DiskPressure,Status:False,LastHeartbeatTime:2022-11-26 05:32:57 +0000 UTC,LastTransitionTime:2022-11-26 05:05:15 +0000 UTC,Reason:KubeletHasNoDiskPressure,Message:kubelet has no disk pressure,},NodeCondition{Type:PIDPressure,Status:False,LastHeartbeatTime:2022-11-26 05:32:57 +0000 UTC,LastTransitionTime:2022-11-26 05:05:15 +0000 UTC,Reason:KubeletHasSufficientPID,Message:kubelet has sufficient PID available,},NodeCondition{Type:Ready,Status:True,LastHeartbeatTime:2022-11-26 05:32:57 +0000 UTC,LastTransitionTime:2022-11-26 05:05:17 +0000 UTC,Reason:KubeletReady,Message:kubelet is posting ready status. AppArmor enabled,},},Addresses:[]NodeAddress{NodeAddress{Type:InternalIP,Address:10.138.0.4,},NodeAddress{Type:ExternalIP,Address:35.230.27.45,},NodeAddress{Type:InternalDNS,Address:bootstrap-e2e-minion-group-rwsn.c.k8s-jkns-e2e-gce-ci-reboot.internal,},NodeAddress{Type:Hostname,Address:bootstrap-e2e-minion-group-rwsn.c.k8s-jkns-e2e-gce-ci-reboot.internal,},},DaemonEndpoints:NodeDaemonEndpoints{KubeletEndpoint:DaemonEndpoint{Port:10250,},},NodeInfo:NodeSystemInfo{MachineID:69e22ce1281686d6cde1b3211b34b8af,SystemUUID:69e22ce1-2816-86d6-cde1-b3211b34b8af,BootID:79e9e1fd-2a39-42ec-b72b-72836d213862,KernelVersion:5.10.123+,OSImage:Container-Optimized OS from Google,ContainerRuntimeVersion:containerd://1.7.0-beta.0-149-gd06318622,KubeletVersion:v1.27.0-alpha.0.50+70617042976dc1,KubeProxyVersion:v1.27.0-alpha.0.50+70617042976dc1,OperatingSystem:linux,Architecture:amd64,},Images:[]ContainerImage{ContainerImage{Names:[registry.k8s.io/e2e-test-images/jessie-dnsutils@sha256:24aaf2626d6b27864c29de2097e8bbb840b3a414271bf7c8995e431e47d8408e registry.k8s.io/e2e-test-images/jessie-dnsutils:1.7],SizeBytes:112030336,},ContainerImage{Names:[registry.k8s.io/e2e-test-images/volume/nfs@sha256:3bda73f2428522b0e342af80a0b9679e8594c2126f2b3cca39ed787589741b9e registry.k8s.io/e2e-test-images/volume/nfs:1.3],SizeBytes:95836203,},ContainerImage{Names:[registry.k8s.io/sig-storage/nfs-provisioner@sha256:e943bb77c7df05ebdc8c7888b2db289b13bf9f012d6a3a5a74f14d4d5743d439 registry.k8s.io/sig-storage/nfs-provisioner:v3.0.1],SizeBytes:90632047,},ContainerImage{Names:[registry.k8s.io/kube-proxy-amd64:v1.27.0-alpha.0.50_70617042976dc1],SizeBytes:67201736,},ContainerImage{Names:[registry.k8s.io/e2e-test-images/agnhost@sha256:16bbf38c463a4223d8cfe4da12bc61010b082a79b4bb003e2d3ba3ece5dd5f9e registry.k8s.io/e2e-test-images/agnhost:2.43],SizeBytes:51706353,},ContainerImage{Names:[gke.gcr.io/prometheus-to-sd@sha256:e739643c3939ba0b161425f45a1989eedfc4a3b166db9a7100863296b4c70510 gke.gcr.io/prometheus-to-sd:v0.11.1-gke.1],SizeBytes:48742566,},ContainerImage{Names:[registry.k8s.io/metrics-server/metrics-server@sha256:6385aec64bb97040a5e692947107b81e178555c7a5b71caa90d733e4130efc10 registry.k8s.io/metrics-server/metrics-server:v0.5.2],SizeBytes:26023008,},ContainerImage{Names:[registry.k8s.io/sig-storage/csi-provisioner@sha256:ee3b525d5b89db99da3b8eb521d9cd90cb6e9ef0fbb651e98bb37be78d36b5b8 registry.k8s.io/sig-storage/csi-provisioner:v3.3.0],SizeBytes:25491225,},ContainerImage{Names:[registry.k8s.io/sig-storage/csi-resizer@sha256:425d8f1b769398127767b06ed97ce62578a3179bcb99809ce93a1649e025ffe7 registry.k8s.io/sig-storage/csi-resizer:v1.6.0],SizeBytes:24148884,},ContainerImage{Names:[registry.k8s.io/sig-storage/csi-snapshotter@sha256:291334908ddf71a4661fd7f6d9d97274de8a5378a2b6fdfeb2ce73414a34f82f registry.k8s.io/sig-storage/csi-snapshotter:v6.1.0],SizeBytes:23881995,},ContainerImage{Names:[registry.k8s.io/sig-storage/csi-attacher@sha256:9a685020911e2725ad019dbce6e4a5ab93d51e3d4557f115e64343345e05781b registry.k8s.io/sig-storage/csi-attacher:v4.0.0],SizeBytes:23847201,},ContainerImage{Names:[registry.k8s.io/sig-storage/snapshot-controller@sha256:823c75d0c45d1427f6d850070956d9ca657140a7bbf828381541d1d808475280 registry.k8s.io/sig-storage/snapshot-controller:v6.1.0],SizeBytes:22620891,},ContainerImage{Names:[registry.k8s.io/sig-storage/hostpathplugin@sha256:92257881c1d6493cf18299a24af42330f891166560047902b8d431fb66b01af5 registry.k8s.io/sig-storage/hostpathplugin:v1.9.0],SizeBytes:18758628,},ContainerImage{Names:[registry.k8s.io/cpa/cluster-proportional-autoscaler@sha256:fd636b33485c7826fb20ef0688a83ee0910317dbb6c0c6f3ad14661c1db25def registry.k8s.io/cpa/cluster-proportional-autoscaler:1.8.4],SizeBytes:15209393,},ContainerImage{Names:[registry.k8s.io/coredns/coredns@sha256:8e352a029d304ca7431c6507b56800636c321cb52289686a581ab70aaa8a2e2a registry.k8s.io/coredns/coredns:v1.9.3],SizeBytes:14837849,},ContainerImage{Names:[registry.k8s.io/autoscaling/addon-resizer@sha256:43f129b81d28f0fdd54de6d8e7eacd5728030782e03db16087fc241ad747d3d6 registry.k8s.io/autoscaling/addon-resizer:1.8.14],SizeBytes:10153852,},ContainerImage{Names:[registry.k8s.io/sig-storage/csi-node-driver-registrar@sha256:0103eee7c35e3e0b5cd8cdca9850dc71c793cdeb6669d8be7a89440da2d06ae4 registry.k8s.io/sig-storage/csi-node-driver-registrar:v2.5.1],SizeBytes:9133109,},ContainerImage{Names:[registry.k8s.io/sig-storage/livenessprobe@sha256:933940f13b3ea0abc62e656c1aa5c5b47c04b15d71250413a6b821bd0c58b94e registry.k8s.io/sig-storage/livenessprobe:v2.7.0],SizeBytes:8688564,},ContainerImage{Names:[registry.k8s.io/kas-network-proxy/proxy-agent@sha256:48f2a4ec3e10553a81b8dd1c6fa5fe4bcc9617f78e71c1ca89c6921335e2d7da registry.k8s.io/kas-network-proxy/proxy-agent:v0.0.33],SizeBytes:8512162,},ContainerImage{Names:[registry.k8s.io/networking/ingress-gce-404-server-with-metrics-amd64@sha256:7eb7b3cee4d33c10c49893ad3c386232b86d4067de5251294d4c620d6e072b93 registry.k8s.io/networking/ingress-gce-404-server-with-metrics-amd64:v1.10.11],SizeBytes:6463068,},ContainerImage{Names:[registry.k8s.io/metadata-proxy@sha256:e914645f22e946bce5165737e1b244e0a296ad1f0f81a9531adc57af2780978a registry.k8s.io/metadata-proxy:v0.1.12],SizeBytes:5301657,},ContainerImage{Names:[registry.k8s.io/e2e-test-images/busybox@sha256:c318242786b139d18676b1c09a0ad7f15fc17f8f16a5b2e625cd0dc8c9703daf registry.k8s.io/e2e-test-images/busybox:1.29-2],SizeBytes:732424,},ContainerImage{Names:[registry.k8s.io/e2e-test-images/busybox@sha256:2e0f836850e09b8b7cc937681d6194537a09fbd5f6b9e08f4d646a85128e8937 registry.k8s.io/e2e-test-images/busybox:1.29-4],SizeBytes:731990,},ContainerImage{Names:[registry.k8s.io/pause@sha256:7031c1b283388d2c2e09b57badb803c05ebed362dc88d84b480cc47f72a21097 registry.k8s.io/pause:3.9],SizeBytes:321520,},ContainerImage{Names:[registry.k8s.io/pause@sha256:9001185023633d17a2f98ff69b6ff2615b8ea02a825adffa40422f51dfdcde9d registry.k8s.io/pause:3.8],SizeBytes:311286,},},VolumesInUse:[kubernetes.io/csi/csi-hostpath-multivolume-4683^6fe7a3e3-6d49-11ed-ac34-8a6e303c97ed kubernetes.io/csi/csi-hostpath-multivolume-4683^7ad9a9cd-6d49-11ed-b2a0-f2f9c395d3b2 kubernetes.io/csi/csi-hostpath-multivolume-6872^8dfb6ea6-6d49-11ed-bbbd-7a3847645704 kubernetes.io/csi/csi-hostpath-provisioning-6984^3b0457df-6d48-11ed-9e4d-d626cdd9a50a],VolumesAttached:[]AttachedVolume{AttachedVolume{Name:kubernetes.io/csi/csi-hostpath-multivolume-6872^8dfb6ea6-6d49-11ed-bbbd-7a3847645704,DevicePath:,},AttachedVolume{Name:kubernetes.io/csi/csi-hostpath-provisioning-6984^3b0457df-6d48-11ed-9e4d-d626cdd9a50a,DevicePath:,},AttachedVolume{Name:kubernetes.io/csi/csi-hostpath-multivolume-4683^7ad9a9cd-6d49-11ed-b2a0-f2f9c395d3b2,DevicePath:,},AttachedVolume{Name:kubernetes.io/csi/csi-hostpath-multivolume-4683^6fe7a3e3-6d49-11ed-ac34-8a6e303c97ed,DevicePath:,},},Config:nil,},} Nov 26 05:34:35.342: INFO: Logging kubelet events for node bootstrap-e2e-minion-group-rwsn Nov 26 05:34:35.404: INFO: Logging pods the kubelet thinks is on node bootstrap-e2e-minion-group-rwsn Nov 26 05:34:35.604: INFO: csi-mockplugin-0 started at 2022-11-26 05:13:01 +0000 UTC (0+4 container statuses recorded) Nov 26 05:34:35.604: INFO: Container busybox ready: false, restart count 6 Nov 26 05:34:35.604: INFO: Container csi-provisioner ready: false, restart count 7 Nov 26 05:34:35.604: INFO: Container driver-registrar ready: false, restart count 7 Nov 26 05:34:35.604: INFO: Container mock ready: false, restart count 7 Nov 26 05:34:35.604: INFO: hostexec-bootstrap-e2e-minion-group-rwsn-ppgcz started at 2022-11-26 05:31:42 +0000 UTC (0+1 container statuses recorded) Nov 26 05:34:35.604: INFO: Container agnhost-container ready: true, restart count 0 Nov 26 05:34:35.604: INFO: hostexec-bootstrap-e2e-minion-group-rwsn-x9txs started at 2022-11-26 05:31:42 +0000 UTC (0+1 container statuses recorded) Nov 26 05:34:35.604: INFO: Container agnhost-container ready: true, restart count 1 Nov 26 05:34:35.604: INFO: csi-mockplugin-0 started at 2022-11-26 05:08:11 +0000 UTC (0+4 container statuses recorded) Nov 26 05:34:35.604: INFO: Container busybox ready: false, restart count 6 Nov 26 05:34:35.604: INFO: Container csi-provisioner ready: false, restart count 8 Nov 26 05:34:35.604: INFO: Container driver-registrar ready: false, restart count 7 Nov 26 05:34:35.604: INFO: Container mock ready: false, restart count 7 Nov 26 05:34:35.604: INFO: csi-hostpathplugin-0 started at 2022-11-26 05:31:42 +0000 UTC (0+7 container statuses recorded) Nov 26 05:34:35.604: INFO: Container csi-attacher ready: true, restart count 1 Nov 26 05:34:35.604: INFO: Container csi-provisioner ready: true, restart count 1 Nov 26 05:34:35.604: INFO: Container csi-resizer ready: true, restart count 1 Nov 26 05:34:35.604: INFO: Container csi-snapshotter ready: true, restart count 1 Nov 26 05:34:35.604: INFO: Container hostpath ready: true, restart count 1 Nov 26 05:34:35.604: INFO: Container liveness-probe ready: true, restart count 1 Nov 26 05:34:35.604: INFO: Container node-driver-registrar ready: true, restart count 1 Nov 26 05:34:35.604: INFO: netserver-1 started at 2022-11-26 05:11:29 +0000 UTC (0+1 container statuses recorded) Nov 26 05:34:35.604: INFO: Container webserver ready: false, restart count 10 Nov 26 05:34:35.604: INFO: l7-default-backend-8549d69d99-8cfzv started at 2022-11-26 05:05:25 +0000 UTC (0+1 container statuses recorded) Nov 26 05:34:35.604: INFO: Container default-http-backend ready: true, restart count 0 Nov 26 05:34:35.604: INFO: volume-snapshot-controller-0 started at 2022-11-26 05:05:25 +0000 UTC (0+1 container statuses recorded) Nov 26 05:34:35.604: INFO: Container volume-snapshot-controller ready: false, restart count 9 Nov 26 05:34:35.604: INFO: netserver-1 started at 2022-11-26 05:09:51 +0000 UTC (0+1 container statuses recorded) Nov 26 05:34:35.604: INFO: Container webserver ready: true, restart count 7 Nov 26 05:34:35.604: INFO: pod-subpath-test-dynamicpv-h6m7 started at 2022-11-26 05:07:31 +0000 UTC (1+1 container statuses recorded) Nov 26 05:34:35.604: INFO: Init container init-volume-dynamicpv-h6m7 ready: true, restart count 0 Nov 26 05:34:35.604: INFO: Container test-container-subpath-dynamicpv-h6m7 ready: false, restart count 0 Nov 26 05:34:35.604: INFO: csi-hostpathplugin-0 started at 2022-11-26 05:16:00 +0000 UTC (0+7 container statuses recorded) Nov 26 05:34:35.604: INFO: Container csi-attacher ready: false, restart count 7 Nov 26 05:34:35.604: INFO: Container csi-provisioner ready: false, restart count 7 Nov 26 05:34:35.604: INFO: Container csi-resizer ready: false, restart count 7 Nov 26 05:34:35.604: INFO: Container csi-snapshotter ready: false, restart count 7 Nov 26 05:34:35.604: INFO: Container hostpath ready: false, restart count 7 Nov 26 05:34:35.604: INFO: Container liveness-probe ready: false, restart count 7 Nov 26 05:34:35.604: INFO: Container node-driver-registrar ready: false, restart count 7 Nov 26 05:34:35.604: INFO: coredns-6d97d5ddb-xk6v4 started at 2022-11-26 05:05:25 +0000 UTC (0+1 container statuses recorded) Nov 26 05:34:35.604: INFO: Container coredns ready: false, restart count 9 Nov 26 05:34:35.604: INFO: pod-04078049-bcc4-43bf-8abc-addcb96042db started at 2022-11-26 05:16:29 +0000 UTC (0+1 container statuses recorded) Nov 26 05:34:35.604: INFO: Container write-pod ready: false, restart count 0 Nov 26 05:34:35.604: INFO: kube-dns-autoscaler-5f6455f985-xktw8 started at 2022-11-26 05:05:25 +0000 UTC (0+1 container statuses recorded) Nov 26 05:34:35.604: INFO: Container autoscaler ready: false, restart count 8 Nov 26 05:34:35.604: INFO: coredns-6d97d5ddb-p2xpt started at 2022-11-26 05:05:31 +0000 UTC (0+1 container statuses recorded) Nov 26 05:34:35.604: INFO: Container coredns ready: false, restart count 9 Nov 26 05:34:35.604: INFO: netserver-1 started at 2022-11-26 05:11:06 +0000 UTC (0+1 container statuses recorded) Nov 26 05:34:35.604: INFO: Container webserver ready: false, restart count 11 Nov 26 05:34:35.604: INFO: pvc-volume-tester-5h2gx started at 2022-11-26 05:13:54 +0000 UTC (0+1 container statuses recorded) Nov 26 05:34:35.604: INFO: Container volume-tester ready: false, restart count 0 Nov 26 05:34:35.604: INFO: kube-proxy-bootstrap-e2e-minion-group-rwsn started at 2022-11-26 05:05:15 +0000 UTC (0+1 container statuses recorded) Nov 26 05:34:35.604: INFO: Container kube-proxy ready: false, restart count 8 Nov 26 05:34:35.604: INFO: csi-hostpathplugin-0 started at 2022-11-26 05:23:21 +0000 UTC (0+7 container statuses recorded) Nov 26 05:34:35.604: INFO: Container csi-attacher ready: false, restart count 6 Nov 26 05:34:35.604: INFO: Container csi-provisioner ready: false, restart count 6 Nov 26 05:34:35.604: INFO: Container csi-resizer ready: false, restart count 6 Nov 26 05:34:35.604: INFO: Container csi-snapshotter ready: false, restart count 6 Nov 26 05:34:35.604: INFO: Container hostpath ready: false, restart count 6 Nov 26 05:34:35.604: INFO: Container liveness-probe ready: false, restart count 6 Nov 26 05:34:35.604: INFO: Container node-driver-registrar ready: false, restart count 6 Nov 26 05:34:35.604: INFO: csi-hostpathplugin-0 started at 2022-11-26 05:16:04 +0000 UTC (0+7 container statuses recorded) Nov 26 05:34:35.604: INFO: Container csi-attacher ready: false, restart count 6 Nov 26 05:34:35.604: INFO: Container csi-provisioner ready: false, restart count 6 Nov 26 05:34:35.604: INFO: Container csi-resizer ready: false, restart count 6 Nov 26 05:34:35.604: INFO: Container csi-snapshotter ready: false, restart count 6 Nov 26 05:34:35.604: INFO: Container hostpath ready: false, restart count 6 Nov 26 05:34:35.604: INFO: Container liveness-probe ready: false, restart count 6 Nov 26 05:34:35.604: INFO: Container node-driver-registrar ready: false, restart count 6 Nov 26 05:34:35.604: INFO: metadata-proxy-v0.1-qlckk started at 2022-11-26 05:05:16 +0000 UTC (0+2 container statuses recorded) Nov 26 05:34:35.604: INFO: Container metadata-proxy ready: true, restart count 0 Nov 26 05:34:35.604: INFO: Container prometheus-to-sd-exporter ready: true, restart count 0 Nov 26 05:34:35.604: INFO: csi-hostpathplugin-0 started at 2022-11-26 05:10:11 +0000 UTC (0+7 container statuses recorded) Nov 26 05:34:35.604: INFO: Container csi-attacher ready: false, restart count 9 Nov 26 05:34:35.604: INFO: Container csi-provisioner ready: false, restart count 9 Nov 26 05:34:35.604: INFO: Container csi-resizer ready: false, restart count 9 Nov 26 05:34:35.604: INFO: Container csi-snapshotter ready: false, restart count 9 Nov 26 05:34:35.604: INFO: Container hostpath ready: false, restart count 9 Nov 26 05:34:35.604: INFO: Container liveness-probe ready: false, restart count 9 Nov 26 05:34:35.604: INFO: Container node-driver-registrar ready: false, restart count 9 Nov 26 05:34:35.604: INFO: csi-mockplugin-0 started at 2022-11-26 05:23:19 +0000 UTC (0+4 container statuses recorded) Nov 26 05:34:35.604: INFO: Container busybox ready: false, restart count 5 Nov 26 05:34:35.604: INFO: Container csi-provisioner ready: false, restart count 4 Nov 26 05:34:35.604: INFO: Container driver-registrar ready: false, restart count 5 Nov 26 05:34:35.604: INFO: Container mock ready: false, restart count 5 Nov 26 05:34:35.604: INFO: konnectivity-agent-sxg5q started at 2022-11-26 05:05:25 +0000 UTC (0+1 container statuses recorded) Nov 26 05:34:35.604: INFO: Container konnectivity-agent ready: false, restart count 8 Nov 26 05:34:36.135: INFO: Latency metrics for node bootstrap-e2e-minion-group-rwsn Nov 26 05:34:36.135: INFO: Logging node info for node bootstrap-e2e-minion-group-vhdj Nov 26 05:34:36.177: INFO: Node Info: &Node{ObjectMeta:{bootstrap-e2e-minion-group-vhdj 0d27fd16-9ec8-4428-bb1e-97fead6c7a66 12181 0 2022-11-26 05:05:23 +0000 UTC <nil> <nil> map[beta.kubernetes.io/arch:amd64 beta.kubernetes.io/instance-type:n1-standard-2 beta.kubernetes.io/os:linux cloud.google.com/metadata-proxy-ready:true failure-domain.beta.kubernetes.io/region:us-west1 failure-domain.beta.kubernetes.io/zone:us-west1-b kubernetes.io/arch:amd64 kubernetes.io/hostname:bootstrap-e2e-minion-group-vhdj kubernetes.io/os:linux node.kubernetes.io/instance-type:n1-standard-2 topology.hostpath.csi/node:bootstrap-e2e-minion-group-vhdj topology.kubernetes.io/region:us-west1 topology.kubernetes.io/zone:us-west1-b] map[csi.volume.kubernetes.io/nodeid:{"csi-hostpath-provisioning-7617":"bootstrap-e2e-minion-group-vhdj","csi-hostpath-volumeio-4786":"bootstrap-e2e-minion-group-vhdj","csi-mock-csi-mock-volumes-5516":"bootstrap-e2e-minion-group-vhdj"} node.alpha.kubernetes.io/ttl:0 volumes.kubernetes.io/controller-managed-attach-detach:true] [] [] [{kubelet Update v1 2022-11-26 05:05:23 +0000 UTC FieldsV1 {"f:metadata":{"f:annotations":{".":{},"f:volumes.kubernetes.io/controller-managed-attach-detach":{}},"f:labels":{".":{},"f:beta.kubernetes.io/arch":{},"f:beta.kubernetes.io/instance-type":{},"f:beta.kubernetes.io/os":{},"f:cloud.google.com/metadata-proxy-ready":{},"f:failure-domain.beta.kubernetes.io/region":{},"f:failure-domain.beta.kubernetes.io/zone":{},"f:kubernetes.io/arch":{},"f:kubernetes.io/hostname":{},"f:kubernetes.io/os":{},"f:node.kubernetes.io/instance-type":{},"f:topology.kubernetes.io/region":{},"f:topology.kubernetes.io/zone":{}}},"f:spec":{"f:providerID":{}}} } {kube-controller-manager Update v1 2022-11-26 05:05:24 +0000 UTC FieldsV1 {"f:metadata":{"f:annotations":{"f:node.alpha.kubernetes.io/ttl":{}}},"f:spec":{"f:podCIDR":{},"f:podCIDRs":{".":{},"v:\"10.64.3.0/24\"":{}}}} } {kube-controller-manager Update v1 2022-11-26 05:17:38 +0000 UTC FieldsV1 {"f:status":{"f:conditions":{"k:{\"type\":\"NetworkUnavailable\"}":{"f:lastHeartbeatTime":{},"f:lastTransitionTime":{},"f:message":{},"f:reason":{},"f:status":{}}}}} status} {node-problem-detector Update v1 2022-11-26 05:30:29 +0000 UTC FieldsV1 {"f:status":{"f:conditions":{"k:{\"type\":\"CorruptDockerOverlay2\"}":{".":{},"f:lastHeartbeatTime":{},"f:lastTransitionTime":{},"f:message":{},"f:reason":{},"f:status":{},"f:type":{}},"k:{\"type\":\"FrequentContainerdRestart\"}":{".":{},"f:lastHeartbeatTime":{},"f:lastTransitionTime":{},"f:message":{},"f:reason":{},"f:status":{},"f:type":{}},"k:{\"type\":\"FrequentDockerRestart\"}":{".":{},"f:lastHeartbeatTime":{},"f:lastTransitionTime":{},"f:message":{},"f:reason":{},"f:status":{},"f:type":{}},"k:{\"type\":\"FrequentKubeletRestart\"}":{".":{},"f:lastHeartbeatTime":{},"f:lastTransitionTime":{},"f:message":{},"f:reason":{},"f:status":{},"f:type":{}},"k:{\"type\":\"FrequentUnregisterNetDevice\"}":{".":{},"f:lastHeartbeatTime":{},"f:lastTransitionTime":{},"f:message":{},"f:reason":{},"f:status":{},"f:type":{}},"k:{\"type\":\"KernelDeadlock\"}":{".":{},"f:lastHeartbeatTime":{},"f:lastTransitionTime":{},"f:message":{},"f:reason":{},"f:status":{},"f:type":{}},"k:{\"type\":\"ReadonlyFilesystem\"}":{".":{},"f:lastHeartbeatTime":{},"f:lastTransitionTime":{},"f:message":{},"f:reason":{},"f:status":{},"f:type":{}}}}} status} {kubelet Update v1 2022-11-26 05:34:01 +0000 UTC FieldsV1 {"f:metadata":{"f:annotations":{"f:csi.volume.kubernetes.io/nodeid":{}},"f:labels":{"f:topology.hostpath.csi/node":{}}},"f:status":{"f:conditions":{"k:{\"type\":\"DiskPressure\"}":{"f:lastHeartbeatTime":{}},"k:{\"type\":\"MemoryPressure\"}":{"f:lastHeartbeatTime":{}},"k:{\"type\":\"PIDPressure\"}":{"f:lastHeartbeatTime":{}},"k:{\"type\":\"Ready\"}":{"f:lastHeartbeatTime":{},"f:lastTransitionTime":{},"f:message":{},"f:reason":{},"f:status":{}}},"f:images":{}}} status}]},Spec:NodeSpec{PodCIDR:10.64.3.0/24,DoNotUseExternalID:,ProviderID:gce://k8s-jkns-e2e-gce-ci-reboot/us-west1-b/bootstrap-e2e-minion-group-vhdj,Unschedulable:false,Taints:[]Taint{},ConfigSource:nil,PodCIDRs:[10.64.3.0/24],},Status:NodeStatus{Capacity:ResourceList{attachable-volumes-gce-pd: {{127 0} {<nil>} 127 DecimalSI},cpu: {{2 0} {<nil>} 2 DecimalSI},ephemeral-storage: {{101203873792 0} {<nil>} 98831908Ki BinarySI},hugepages-1Gi: {{0 0} {<nil>} 0 DecimalSI},hugepages-2Mi: {{0 0} {<nil>} 0 DecimalSI},memory: {{7815438336 0} {<nil>} BinarySI},pods: {{110 0} {<nil>} 110 DecimalSI},},Allocatable:ResourceList{attachable-volumes-gce-pd: {{127 0} {<nil>} 127 DecimalSI},cpu: {{2 0} {<nil>} 2 DecimalSI},ephemeral-storage: {{91083486262 0} {<nil>} 91083486262 DecimalSI},hugepages-1Gi: {{0 0} {<nil>} 0 DecimalSI},hugepages-2Mi: {{0 0} {<nil>} 0 DecimalSI},memory: {{7553294336 0} {<nil>} BinarySI},pods: {{110 0} {<nil>} 110 DecimalSI},},Phase:,Conditions:[]NodeCondition{NodeCondition{Type:KernelDeadlock,Status:False,LastHeartbeatTime:2022-11-26 05:30:29 +0000 UTC,LastTransitionTime:2022-11-26 05:05:26 +0000 UTC,Reason:KernelHasNoDeadlock,Message:kernel has no deadlock,},NodeCondition{Type:ReadonlyFilesystem,Status:False,LastHeartbeatTime:2022-11-26 05:30:29 +0000 UTC,LastTransitionTime:2022-11-26 05:05:26 +0000 UTC,Reason:FilesystemIsNotReadOnly,Message:Filesystem is not read-only,},NodeCondition{Type:CorruptDockerOverlay2,Status:False,LastHeartbeatTime:2022-11-26 05:30:29 +0000 UTC,LastTransitionTime:2022-11-26 05:05:26 +0000 UTC,Reason:NoCorruptDockerOverlay2,Message:docker overlay2 is functioning properly,},NodeCondition{Type:FrequentUnregisterNetDevice,Status:False,LastHeartbeatTime:2022-11-26 05:30:29 +0000 UTC,LastTransitionTime:2022-11-26 05:05:26 +0000 UTC,Reason:NoFrequentUnregisterNetDevice,Message:node is functioning properly,},NodeCondition{Type:FrequentKubeletRestart,Status:False,LastHeartbeatTime:2022-11-26 05:30:29 +0000 UTC,LastTransitionTime:2022-11-26 05:05:26 +0000 UTC,Reason:NoFrequentKubeletRestart,Message:kubelet is functioning properly,},NodeCondition{Type:FrequentDockerRestart,Status:False,LastHeartbeatTime:2022-11-26 05:30:29 +0000 UTC,LastTransitionTime:2022-11-26 05:05:26 +0000 UTC,Reason:NoFrequentDockerRestart,Message:docker is functioning properly,},NodeCondition{Type:FrequentContainerdRestart,Status:False,LastHeartbeatTime:2022-11-26 05:30:29 +0000 UTC,LastTransitionTime:2022-11-26 05:05:26 +0000 UTC,Reason:NoFrequentContainerdRestart,Message:containerd is functioning properly,},NodeCondition{Type:NetworkUnavailable,Status:False,LastHeartbeatTime:2022-11-26 05:05:39 +0000 UTC,LastTransitionTime:2022-11-26 05:05:39 +0000 UTC,Reason:RouteCreated,Message:RouteController created a route,},NodeCondition{Type:MemoryPressure,Status:False,LastHeartbeatTime:2022-11-26 05:32:57 +0000 UTC,LastTransitionTime:2022-11-26 05:05:23 +0000 UTC,Reason:KubeletHasSufficientMemory,Message:kubelet has sufficient memory available,},NodeCondition{Type:DiskPressure,Status:False,LastHeartbeatTime:2022-11-26 05:32:57 +0000 UTC,LastTransitionTime:2022-11-26 05:05:23 +0000 UTC,Reason:KubeletHasNoDiskPressure,Message:kubelet has no disk pressure,},NodeCondition{Type:PIDPressure,Status:False,LastHeartbeatTime:2022-11-26 05:32:57 +0000 UTC,LastTransitionTime:2022-11-26 05:05:23 +0000 UTC,Reason:KubeletHasSufficientPID,Message:kubelet has sufficient PID available,},NodeCondition{Type:Ready,Status:True,LastHeartbeatTime:2022-11-26 05:32:57 +0000 UTC,LastTransitionTime:2022-11-26 05:05:24 +0000 UTC,Reason:KubeletReady,Message:kubelet is posting ready status. AppArmor enabled,},},Addresses:[]NodeAddress{NodeAddress{Type:InternalIP,Address:10.138.0.3,},NodeAddress{Type:ExternalIP,Address:34.83.94.215,},NodeAddress{Type:InternalDNS,Address:bootstrap-e2e-minion-group-vhdj.c.k8s-jkns-e2e-gce-ci-reboot.internal,},NodeAddress{Type:Hostname,Address:bootstrap-e2e-minion-group-vhdj.c.k8s-jkns-e2e-gce-ci-reboot.internal,},},DaemonEndpoints:NodeDaemonEndpoints{KubeletEndpoint:DaemonEndpoint{Port:10250,},},NodeInfo:NodeSystemInfo{MachineID:31b0953ad3e740aeb7fc1a89edb5435a,SystemUUID:31b0953a-d3e7-40ae-b7fc-1a89edb5435a,BootID:14c054d0-bda9-4ddf-b9f6-e1742be33d25,KernelVersion:5.10.123+,OSImage:Container-Optimized OS from Google,ContainerRuntimeVersion:containerd://1.7.0-beta.0-149-gd06318622,KubeletVersion:v1.27.0-alpha.0.50+70617042976dc1,KubeProxyVersion:v1.27.0-alpha.0.50+70617042976dc1,OperatingSystem:linux,Architecture:amd64,},Images:[]ContainerImage{ContainerImage{Names:[registry.k8s.io/e2e-test-images/jessie-dnsutils@sha256:24aaf2626d6b27864c29de2097e8bbb840b3a414271bf7c8995e431e47d8408e registry.k8s.io/e2e-test-images/jessie-dnsutils:1.7],SizeBytes:112030336,},ContainerImage{Names:[registry.k8s.io/sig-storage/nfs-provisioner@sha256:e943bb77c7df05ebdc8c7888b2db289b13bf9f012d6a3a5a74f14d4d5743d439 registry.k8s.io/sig-storage/nfs-provisioner:v3.0.1],SizeBytes:90632047,},ContainerImage{Names:[registry.k8s.io/kube-proxy-amd64:v1.27.0-alpha.0.50_70617042976dc1],SizeBytes:67201736,},ContainerImage{Names:[registry.k8s.io/e2e-test-images/agnhost@sha256:16bbf38c463a4223d8cfe4da12bc61010b082a79b4bb003e2d3ba3ece5dd5f9e registry.k8s.io/e2e-test-images/agnhost:2.43],SizeBytes:51706353,},ContainerImage{Names:[gke.gcr.io/prometheus-to-sd@sha256:e739643c3939ba0b161425f45a1989eedfc4a3b166db9a7100863296b4c70510 gke.gcr.io/prometheus-to-sd:v0.11.1-gke.1],SizeBytes:48742566,},ContainerImage{Names:[registry.k8s.io/e2e-test-images/httpd@sha256:148b022f5c5da426fc2f3c14b5c0867e58ef05961510c84749ac1fddcb0fef22 registry.k8s.io/e2e-test-images/httpd:2.4.38-4],SizeBytes:40764257,},ContainerImage{Names:[registry.k8s.io/sig-storage/csi-provisioner@sha256:ee3b525d5b89db99da3b8eb521d9cd90cb6e9ef0fbb651e98bb37be78d36b5b8 registry.k8s.io/sig-storage/csi-provisioner:v3.3.0],SizeBytes:25491225,},ContainerImage{Names:[registry.k8s.io/sig-storage/csi-resizer@sha256:425d8f1b769398127767b06ed97ce62578a3179bcb99809ce93a1649e025ffe7 registry.k8s.io/sig-storage/csi-resizer:v1.6.0],SizeBytes:24148884,},ContainerImage{Names:[registry.k8s.io/sig-storage/csi-snapshotter@sha256:291334908ddf71a4661fd7f6d9d97274de8a5378a2b6fdfeb2ce73414a34f82f registry.k8s.io/sig-storage/csi-snapshotter:v6.1.0],SizeBytes:23881995,},ContainerImage{Names:[registry.k8s.io/sig-storage/csi-attacher@sha256:9a685020911e2725ad019dbce6e4a5ab93d51e3d4557f115e64343345e05781b registry.k8s.io/sig-storage/csi-attacher:v4.0.0],SizeBytes:23847201,},ContainerImage{Names:[registry.k8s.io/sig-storage/hostpathplugin@sha256:92257881c1d6493cf18299a24af42330f891166560047902b8d431fb66b01af5 registry.k8s.io/sig-storage/hostpathplugin:v1.9.0],SizeBytes:18758628,},ContainerImage{Names:[registry.k8s.io/sig-storage/csi-node-driver-registrar@sha256:0103eee7c35e3e0b5cd8cdca9850dc71c793cdeb6669d8be7a89440da2d06ae4 registry.k8s.io/sig-storage/csi-node-driver-registrar:v2.5.1],SizeBytes:9133109,},ContainerImage{Names:[registry.k8s.io/sig-storage/livenessprobe@sha256:933940f13b3ea0abc62e656c1aa5c5b47c04b15d71250413a6b821bd0c58b94e registry.k8s.io/sig-storage/livenessprobe:v2.7.0],SizeBytes:8688564,},ContainerImage{Names:[registry.k8s.io/kas-network-proxy/proxy-agent@sha256:48f2a4ec3e10553a81b8dd1c6fa5fe4bcc9617f78e71c1ca89c6921335e2d7da registry.k8s.io/kas-network-proxy/proxy-agent:v0.0.33],SizeBytes:8512162,},ContainerImage{Names:[registry.k8s.io/metadata-proxy@sha256:e914645f22e946bce5165737e1b244e0a296ad1f0f81a9531adc57af2780978a registry.k8s.io/metadata-proxy:v0.1.12],SizeBytes:5301657,},ContainerImage{Names:[registry.k8s.io/e2e-test-images/busybox@sha256:2e0f836850e09b8b7cc937681d6194537a09fbd5f6b9e08f4d646a85128e8937 registry.k8s.io/e2e-test-images/busybox:1.29-4],SizeBytes:731990,},ContainerImage{Names:[registry.k8s.io/pause@sha256:9001185023633d17a2f98ff69b6ff2615b8ea02a825adffa40422f51dfdcde9d registry.k8s.io/pause:3.8],SizeBytes:311286,},},VolumesInUse:[],VolumesAttached:[]AttachedVolume{},Config:nil,},} Nov 26 05:34:36.177: INFO: Logging kubelet events for node bootstrap-e2e-minion-group-vhdj Nov 26 05:34:36.273: INFO: Logging pods the kubelet thinks is on node bootstrap-e2e-minion-group-vhdj Nov 26 05:34:36.467: INFO: test-hostpath-type-d4l5p started at 2022-11-26 05:31:42 +0000 UTC (0+1 container statuses recorded) Nov 26 05:34:36.467: INFO: Container host-path-sh-testing ready: true, restart count 0 Nov 26 05:34:36.467: INFO: netserver-2 started at 2022-11-26 05:09:51 +0000 UTC (0+1 container statuses recorded) Nov 26 05:34:36.467: INFO: Container webserver ready: false, restart count 5 Nov 26 05:34:36.467: INFO: netserver-2 started at 2022-11-26 05:11:29 +0000 UTC (0+1 container statuses recorded) Nov 26 05:34:36.467: INFO: Container webserver ready: true, restart count 3 Nov 26 05:34:36.467: INFO: hostexec-bootstrap-e2e-minion-group-vhdj-2hxht started at 2022-11-26 05:31:42 +0000 UTC (0+1 container statuses recorded) Nov 26 05:34:36.467: INFO: Container agnhost-container ready: true, restart count 3 Nov 26 05:34:36.467: INFO: netserver-2 started at 2022-11-26 05:11:06 +0000 UTC (0+1 container statuses recorded) Nov 26 05:34:36.467: INFO: Container webserver ready: true, restart count 2 Nov 26 05:34:36.467: INFO: csi-mockplugin-0 started at 2022-11-26 05:31:42 +0000 UTC (0+3 container statuses recorded) Nov 26 05:34:36.467: INFO: Container csi-provisioner ready: true, restart count 0 Nov 26 05:34:36.467: INFO: Container driver-registrar ready: true, restart count 0 Nov 26 05:34:36.467: INFO: Container mock ready: true, restart count 0 Nov 26 05:34:36.467: INFO: forbid-27824009-xr9r6 started at 2022-11-26 05:31:42 +0000 UTC (0+1 container statuses recorded) Nov 26 05:34:36.467: INFO: Container c ready: true, restart count 2 Nov 26 05:34:36.467: INFO: kube-proxy-bootstrap-e2e-minion-group-vhdj started at 2022-11-26 05:05:23 +0000 UTC (0+1 container statuses recorded) Nov 26 05:34:36.467: INFO: Container kube-proxy ready: true, restart count 9 Nov 26 05:34:36.467: INFO: pod-9fe7e984-9a6a-4bb0-9c65-0628a6a2f792 started at 2022-11-26 05:08:13 +0000 UTC (0+1 container statuses recorded) Nov 26 05:34:36.467: INFO: Container write-pod ready: false, restart count 0 Nov 26 05:34:36.467: INFO: external-provisioner-zp7k7 started at 2022-11-26 05:31:42 +0000 UTC (0+1 container statuses recorded) Nov 26 05:34:36.467: INFO: Container nfs-provisioner ready: true, restart count 0 Nov 26 05:34:36.467: INFO: mutability-test-7ktz4 started at 2022-11-26 05:23:40 +0000 UTC (0+1 container statuses recorded) Nov 26 05:34:36.467: INFO: Container netexec ready: false, restart count 5 Nov 26 05:34:36.467: INFO: metadata-proxy-v0.1-wn5l5 started at 2022-11-26 05:05:24 +0000 UTC (0+2 container statuses recorded) Nov 26 05:34:36.467: INFO: Container metadata-proxy ready: true, restart count 0 Nov 26 05:34:36.467: INFO: Container prometheus-to-sd-exporter ready: true, restart count 0 Nov 26 05:34:36.467: INFO: pod-subpath-test-dynamicpv-4nfp started at 2022-11-26 05:11:17 +0000 UTC (1+2 container statuses recorded) Nov 26 05:34:36.467: INFO: Init container init-volume-dynamicpv-4nfp ready: true, restart count 0 Nov 26 05:34:36.467: INFO: Container test-container-subpath-dynamicpv-4nfp ready: false, restart count 2 Nov 26 05:34:36.467: INFO: Container test-container-volume-dynamicpv-4nfp ready: false, restart count 2 Nov 26 05:34:36.467: INFO: test-container-pod started at 2022-11-26 05:12:10 +0000 UTC (0+1 container statuses recorded) Nov 26 05:34:36.467: INFO: Container webserver ready: true, restart count 6 Nov 26 05:34:36.467: INFO: test-hostpath-type-b42b4 started at 2022-11-26 05:31:42 +0000 UTC (0+1 container statuses recorded) Nov 26 05:34:36.467: INFO: Container host-path-testing ready: true, restart count 0 Nov 26 05:34:36.467: INFO: csi-hostpathplugin-0 started at 2022-11-26 05:13:07 +0000 UTC (0+7 container statuses recorded) Nov 26 05:34:36.467: INFO: Container csi-attacher ready: true, restart count 5 Nov 26 05:34:36.467: INFO: Container csi-provisioner ready: true, restart count 5 Nov 26 05:34:36.467: INFO: Container csi-resizer ready: true, restart count 5 Nov 26 05:34:36.467: INFO: Container csi-snapshotter ready: true, restart count 5 Nov 26 05:34:36.467: INFO: Container hostpath ready: true, restart count 5 Nov 26 05:34:36.467: INFO: Container liveness-probe ready: true, restart count 5 Nov 26 05:34:36.467: INFO: Container node-driver-registrar ready: true, restart count 5 Nov 26 05:34:36.467: INFO: hostpathsymlink-io-client started at 2022-11-26 05:31:42 +0000 UTC (1+1 container statuses recorded) Nov 26 05:34:36.467: INFO: Init container hostpathsymlink-io-init ready: true, restart count 0 Nov 26 05:34:36.467: INFO: Container hostpathsymlink-io-client ready: false, restart count 0 Nov 26 05:34:36.467: INFO: konnectivity-agent-nt4fd started at 2022-11-26 05:05:39 +0000 UTC (0+1 container statuses recorded) Nov 26 05:34:36.467: INFO: Container konnectivity-agent ready: false, restart count 8 Nov 26 05:34:36.467: INFO: csi-hostpathplugin-0 started at 2022-11-26 05:10:50 +0000 UTC (0+7 container statuses recorded) Nov 26 05:34:36.467: INFO: Container csi-attacher ready: true, restart count 6 Nov 26 05:34:36.467: INFO: Container csi-provisioner ready: true, restart count 6 Nov 26 05:34:36.467: INFO: Container csi-resizer ready: true, restart count 6 Nov 26 05:34:36.467: INFO: Container csi-snapshotter ready: true, restart count 6 Nov 26 05:34:36.467: INFO: Container hostpath ready: true, restart count 6 Nov 26 05:34:36.467: INFO: Container liveness-probe ready: true, restart count 6 Nov 26 05:34:36.467: INFO: Container node-driver-registrar ready: true, restart count 6 Nov 26 05:34:36.467: INFO: test-hostpath-type-w4sl2 started at 2022-11-26 05:31:42 +0000 UTC (0+1 container statuses recorded) Nov 26 05:34:36.467: INFO: Container host-path-sh-testing ready: false, restart count 0 Nov 26 05:34:36.467: INFO: addon-reconcile-test-8rfrt started at 2022-11-26 05:31:42 +0000 UTC (0+1 container statuses recorded) Nov 26 05:34:36.467: INFO: Container addon-reconcile-test ready: true, restart count 3 Nov 26 05:34:36.467: INFO: mutability-test-55dcq started at 2022-11-26 05:13:49 +0000 UTC (0+1 container statuses recorded) Nov 26 05:34:36.467: INFO: Container netexec ready: true, restart count 5 Nov 26 05:34:36.467: INFO: hostexec-bootstrap-e2e-minion-group-vhdj-7m4fs started at 2022-11-26 05:31:42 +0000 UTC (0+1 container statuses recorded) Nov 26 05:34:36.467: INFO: Container agnhost-container ready: true, restart count 0 Nov 26 05:34:37.462: INFO: Latency metrics for node bootstrap-e2e-minion-group-vhdj [DeferCleanup (Each)] [sig-apps] StatefulSet tear down framework | framework.go:193 STEP: Destroying namespace "statefulset-3690" for this suite. 11/26/22 05:34:37.462
Filter through log files | View test history on testgrid
go run hack/e2e.go -v --test --test_args='--ginkgo.focus=Kubernetes\se2e\ssuite\s\[It\]\s\[sig\-auth\]\sServiceAccounts\sshould\ssupport\sInClusterConfig\swith\stoken\srotation\s\[Slow\]$'
test/e2e/auth/service_accounts.go:520 k8s.io/kubernetes/test/e2e/auth.glob..func5.6() test/e2e/auth/service_accounts.go:520 +0x9abfrom junit_01.xml
[BeforeEach] [sig-auth] ServiceAccounts set up framework | framework.go:178 STEP: Creating a kubernetes client 11/26/22 05:12:36.111 Nov 26 05:12:36.111: INFO: >>> kubeConfig: /workspace/.kube/config STEP: Building a namespace api object, basename svcaccounts 11/26/22 05:12:36.113 STEP: Waiting for a default service account to be provisioned in namespace 11/26/22 05:12:36.299 STEP: Waiting for kube-root-ca.crt to be provisioned in namespace 11/26/22 05:12:36.426 [BeforeEach] [sig-auth] ServiceAccounts test/e2e/framework/metrics/init/init.go:31 [It] should support InClusterConfig with token rotation [Slow] test/e2e/auth/service_accounts.go:432 Nov 26 05:12:36.745: INFO: created pod Nov 26 05:12:36.745: INFO: Waiting up to 1m0s for 1 pods to be running and ready: [inclusterclient] Nov 26 05:12:36.745: INFO: Waiting up to 1m0s for pod "inclusterclient" in namespace "svcaccounts-6307" to be "running and ready" Nov 26 05:12:36.819: INFO: Pod "inclusterclient": Phase="Pending", Reason="", readiness=false. Elapsed: 74.068215ms Nov 26 05:12:36.819: INFO: Error evaluating pod condition running and ready: want pod 'inclusterclient' on 'bootstrap-e2e-minion-group-0975' to be 'Running' but was 'Pending' Nov 26 05:12:38.981: INFO: Pod "inclusterclient": Phase="Running", Reason="", readiness=true. Elapsed: 2.236111806s Nov 26 05:12:38.981: INFO: Pod "inclusterclient" satisfied condition "running and ready" Nov 26 05:12:38.981: INFO: Wanted all 1 pods to be running and ready. Result: true. Pods: [inclusterclient] Nov 26 05:12:38.981: INFO: pod is ready Nov 26 05:13:38.981: INFO: polling logs Nov 26 05:13:39.085: INFO: Retrying. Still waiting to see more unique tokens: got=1, want=2 Nov 26 05:14:38.982: INFO: polling logs Nov 26 05:14:46.529: INFO: Error pulling logs: an error on the server ("unknown") has prevented the request from succeeding (get pods inclusterclient) Nov 26 05:15:38.982: INFO: polling logs Nov 26 05:15:39.039: FAIL: Unexpected error: inclusterclient reported an error: saw status=failed I1126 05:12:37.731704 1 main.go:61] started I1126 05:13:07.736070 1 main.go:79] calling /healthz I1126 05:13:07.742235 1 main.go:96] authz_header=08NWHZC_7WcGS2S4c_q2G0RE8Q1k6tOcFW0eHg5oWkM I1126 05:13:37.735051 1 main.go:79] calling /healthz I1126 05:13:37.735306 1 main.go:96] authz_header=08NWHZC_7WcGS2S4c_q2G0RE8Q1k6tOcFW0eHg5oWkM I1126 05:14:07.736219 1 main.go:79] calling /healthz I1126 05:14:07.736613 1 main.go:96] authz_header=08NWHZC_7WcGS2S4c_q2G0RE8Q1k6tOcFW0eHg5oWkM E1126 05:14:07.737500 1 main.go:82] status=failed E1126 05:14:07.737520 1 main.go:83] error checking /healthz: Get "https://10.0.0.1:443/healthz": dial tcp 10.0.0.1:443: connect: connection refused I1126 05:14:37.733669 1 main.go:79] calling /healthz I1126 05:14:37.733944 1 main.go:96] authz_header=08NWHZC_7WcGS2S4c_q2G0RE8Q1k6tOcFW0eHg5oWkM E1126 05:14:44.134239 1 main.go:82] status=failed E1126 05:14:44.134288 1 main.go:83] error checking /healthz: an error on the server ("[+]ping ok\n[+]log ok\n[+]etcd ok\n[+]poststarthook/start-kube-apiserver-admission-initializer ok\n[+]poststarthook/generic-apiserver-start-informers ok\n[+]poststarthook/priority-and-fairness-config-consumer ok\n[+]poststarthook/priority-and-fairness-filter ok\n[+]poststarthook/storage-object-count-tracker-hook ok\n[+]poststarthook/start-apiextensions-informers ok\n[-]poststarthook/start-apiextensions-controllers failed: reason withheld\n[-]poststarthook/crd-informer-synced failed: reason withheld\n[-]poststarthook/bootstrap-controller failed: reason withheld\n[-]poststarthook/rbac/bootstrap-roles failed: reason withheld\n[-]poststarthook/scheduling/bootstrap-system-priority-classes failed: reason withheld\n[-]poststarthook/priority-and-fairness-config-producer failed: reason withheld\n[+]poststarthook/start-cluster-authentication-info-controller ok\n[+]poststarthook/start-kube-apiserver-identity-lease-controller ok\n[+]poststarthook/start-kube-apiserver-identity-lease-garbage-collector ok\n[+]poststarthook/start-legacy-token-tracking-controller ok\n[+]poststarthook/aggregator-reload-proxy-client-cert ok\n[+]poststarthook/start-kube-aggregator-informers ok\n[-]poststarthook/apiservice-registration-controller failed: reason withheld\n[+]poststarthook/apiservice-status-available-controller ok\n[+]poststarthook/kube-apiserver-autoregistration ok\n[+]autoregister-completion ok\n[+]poststarthook/apiservice-openapi-controller ok\n[+]poststarthook/apiservice-openapiv3-controller ok\nhealthz check failed") has prevented the request from succeeding [+]ping ok [+]log ok [+]etcd ok [+]poststarthook/start-kube-apiserver-admission-initializer ok [+]poststarthook/generic-apiserver-start-informers ok [+]poststarthook/priority-and-fairness-config-consumer ok [+]poststarthook/priority-and-fairness-filter ok [+]poststarthook/storage-object-count-tracker-hook ok [+]poststarthook/start-apiextensions-informers ok [-]poststarthook/start-apiextensions-controllers failed: reason withheld [-]poststarthook/crd-informer-synced failed: reason withheld [-]poststarthook/bootstrap-controller failed: reason withheld [-]poststarthook/rbac/bootstrap-roles failed: reason withheld [-]poststarthook/scheduling/bootstrap-system-priority-classes failed: reason withheld [-]poststarthook/priority-and-fairness-config-producer failed: reason withheld [+]poststarthook/start-cluster-authentication-info-controller ok [+]poststarthook/start-kube-apiserver-identity-lease-controller ok [+]poststarthook/start-kube-apiserver-identity-lease-garbage-collector ok [+]poststarthook/start-legacy-token-tracking-controller ok [+]poststarthook/aggregator-reload-proxy-client-cert ok [+]poststarthook/start-kube-aggregator-informers ok [-]poststarthook/apiservice-registration-controller failed: reason withheld [+]poststarthook/apiservice-status-available-controller ok [+]poststarthook/kube-apiserver-autoregistration ok [+]autoregister-completion ok [+]poststarthook/apiservice-openapi-controller ok [+]poststarthook/apiservice-openapiv3-controller ok healthz check failed I1126 05:15:07.732999 1 main.go:79] calling /healthz I1126 05:15:07.733263 1 main.go:96] authz_header=08NWHZC_7WcGS2S4c_q2G0RE8Q1k6tOcFW0eHg5oWkM Full Stack Trace k8s.io/kubernetes/test/e2e/auth.glob..func5.6() test/e2e/auth/service_accounts.go:520 +0x9ab [AfterEach] [sig-auth] ServiceAccounts test/e2e/framework/node/init/init.go:32 Nov 26 05:15:39.039: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready [DeferCleanup (Each)] [sig-auth] ServiceAccounts test/e2e/framework/metrics/init/init.go:33 [DeferCleanup (Each)] [sig-auth] ServiceAccounts dump namespaces | framework.go:196 STEP: dump namespace information after failure 11/26/22 05:15:39.121 STEP: Collecting events from namespace "svcaccounts-6307". 11/26/22 05:15:39.121 STEP: Found 5 events. 11/26/22 05:15:39.162 Nov 26 05:15:39.162: INFO: At 0001-01-01 00:00:00 +0000 UTC - event for inclusterclient: { } Scheduled: Successfully assigned svcaccounts-6307/inclusterclient to bootstrap-e2e-minion-group-0975 Nov 26 05:15:39.162: INFO: At 2022-11-26 05:12:37 +0000 UTC - event for inclusterclient: {kubelet bootstrap-e2e-minion-group-0975} Pulled: Container image "registry.k8s.io/e2e-test-images/agnhost:2.43" already present on machine Nov 26 05:15:39.162: INFO: At 2022-11-26 05:12:37 +0000 UTC - event for inclusterclient: {kubelet bootstrap-e2e-minion-group-0975} Created: Created container inclusterclient Nov 26 05:15:39.162: INFO: At 2022-11-26 05:12:37 +0000 UTC - event for inclusterclient: {kubelet bootstrap-e2e-minion-group-0975} Started: Started container inclusterclient Nov 26 05:15:39.162: INFO: At 2022-11-26 05:15:20 +0000 UTC - event for inclusterclient: {kubelet bootstrap-e2e-minion-group-0975} Killing: Stopping container inclusterclient Nov 26 05:15:39.203: INFO: POD NODE PHASE GRACE CONDITIONS Nov 26 05:15:39.203: INFO: inclusterclient bootstrap-e2e-minion-group-0975 Failed [{Initialized True 0001-01-01 00:00:00 +0000 UTC 2022-11-26 05:12:36 +0000 UTC } {Ready False 0001-01-01 00:00:00 +0000 UTC 2022-11-26 05:15:20 +0000 UTC PodFailed } {ContainersReady False 0001-01-01 00:00:00 +0000 UTC 2022-11-26 05:15:20 +0000 UTC PodFailed } {PodScheduled True 0001-01-01 00:00:00 +0000 UTC 2022-11-26 05:12:36 +0000 UTC }] Nov 26 05:15:39.203: INFO: Nov 26 05:15:39.293: INFO: Logging node info for node bootstrap-e2e-master Nov 26 05:15:39.335: INFO: Node Info: &Node{ObjectMeta:{bootstrap-e2e-master 9a48eb11-c964-4bf8-82c0-1c32a9703981 4609 0 2022-11-26 05:05:18 +0000 UTC <nil> <nil> map[beta.kubernetes.io/arch:amd64 beta.kubernetes.io/instance-type:n1-standard-1 beta.kubernetes.io/os:linux cloud.google.com/metadata-proxy-ready:true failure-domain.beta.kubernetes.io/region:us-west1 failure-domain.beta.kubernetes.io/zone:us-west1-b kubernetes.io/arch:amd64 kubernetes.io/hostname:bootstrap-e2e-master kubernetes.io/os:linux node.kubernetes.io/instance-type:n1-standard-1 topology.kubernetes.io/region:us-west1 topology.kubernetes.io/zone:us-west1-b] map[node.alpha.kubernetes.io/ttl:0 volumes.kubernetes.io/controller-managed-attach-detach:true] [] [] [{kubelet Update v1 2022-11-26 05:05:18 +0000 UTC FieldsV1 {"f:metadata":{"f:annotations":{".":{},"f:volumes.kubernetes.io/controller-managed-attach-detach":{}},"f:labels":{".":{},"f:beta.kubernetes.io/arch":{},"f:beta.kubernetes.io/instance-type":{},"f:beta.kubernetes.io/os":{},"f:cloud.google.com/metadata-proxy-ready":{},"f:failure-domain.beta.kubernetes.io/region":{},"f:failure-domain.beta.kubernetes.io/zone":{},"f:kubernetes.io/arch":{},"f:kubernetes.io/hostname":{},"f:kubernetes.io/os":{},"f:node.kubernetes.io/instance-type":{},"f:topology.kubernetes.io/region":{},"f:topology.kubernetes.io/zone":{}}},"f:spec":{"f:providerID":{},"f:unschedulable":{}}} } {kube-controller-manager Update v1 2022-11-26 05:05:25 +0000 UTC FieldsV1 {"f:status":{"f:conditions":{"k:{\"type\":\"NetworkUnavailable\"}":{"f:lastHeartbeatTime":{},"f:lastTransitionTime":{},"f:message":{},"f:reason":{},"f:status":{}}}}} status} {kube-controller-manager Update v1 2022-11-26 05:05:39 +0000 UTC FieldsV1 {"f:metadata":{"f:annotations":{"f:node.alpha.kubernetes.io/ttl":{}}},"f:spec":{"f:podCIDR":{},"f:podCIDRs":{".":{},"v:\"10.64.1.0/24\"":{}},"f:taints":{}}} } {kubelet Update v1 2022-11-26 05:11:14 +0000 UTC FieldsV1 {"f:status":{"f:conditions":{"k:{\"type\":\"DiskPressure\"}":{"f:lastHeartbeatTime":{}},"k:{\"type\":\"MemoryPressure\"}":{"f:lastHeartbeatTime":{}},"k:{\"type\":\"PIDPressure\"}":{"f:lastHeartbeatTime":{}},"k:{\"type\":\"Ready\"}":{"f:lastHeartbeatTime":{},"f:lastTransitionTime":{},"f:message":{},"f:reason":{},"f:status":{}}},"f:images":{}}} status}]},Spec:NodeSpec{PodCIDR:10.64.1.0/24,DoNotUseExternalID:,ProviderID:gce://k8s-jkns-e2e-gce-ci-reboot/us-west1-b/bootstrap-e2e-master,Unschedulable:true,Taints:[]Taint{Taint{Key:node-role.kubernetes.io/master,Value:,Effect:NoSchedule,TimeAdded:<nil>,},Taint{Key:node.kubernetes.io/unschedulable,Value:,Effect:NoSchedule,TimeAdded:<nil>,},},ConfigSource:nil,PodCIDRs:[10.64.1.0/24],},Status:NodeStatus{Capacity:ResourceList{attachable-volumes-gce-pd: {{127 0} {<nil>} 127 DecimalSI},cpu: {{1 0} {<nil>} 1 DecimalSI},ephemeral-storage: {{16656896000 0} {<nil>} 16266500Ki BinarySI},hugepages-1Gi: {{0 0} {<nil>} 0 DecimalSI},hugepages-2Mi: {{0 0} {<nil>} 0 DecimalSI},memory: {{3858374656 0} {<nil>} BinarySI},pods: {{110 0} {<nil>} 110 DecimalSI},},Allocatable:ResourceList{attachable-volumes-gce-pd: {{127 0} {<nil>} 127 DecimalSI},cpu: {{1 0} {<nil>} 1 DecimalSI},ephemeral-storage: {{14991206376 0} {<nil>} 14991206376 DecimalSI},hugepages-1Gi: {{0 0} {<nil>} 0 DecimalSI},hugepages-2Mi: {{0 0} {<nil>} 0 DecimalSI},memory: {{3596230656 0} {<nil>} BinarySI},pods: {{110 0} {<nil>} 110 DecimalSI},},Phase:,Conditions:[]NodeCondition{NodeCondition{Type:NetworkUnavailable,Status:False,LastHeartbeatTime:2022-11-26 05:05:25 +0000 UTC,LastTransitionTime:2022-11-26 05:05:25 +0000 UTC,Reason:RouteCreated,Message:RouteController created a route,},NodeCondition{Type:MemoryPressure,Status:False,LastHeartbeatTime:2022-11-26 05:11:14 +0000 UTC,LastTransitionTime:2022-11-26 05:05:18 +0000 UTC,Reason:KubeletHasSufficientMemory,Message:kubelet has sufficient memory available,},NodeCondition{Type:DiskPressure,Status:False,LastHeartbeatTime:2022-11-26 05:11:14 +0000 UTC,LastTransitionTime:2022-11-26 05:05:18 +0000 UTC,Reason:KubeletHasNoDiskPressure,Message:kubelet has no disk pressure,},NodeCondition{Type:PIDPressure,Status:False,LastHeartbeatTime:2022-11-26 05:11:14 +0000 UTC,LastTransitionTime:2022-11-26 05:05:18 +0000 UTC,Reason:KubeletHasSufficientPID,Message:kubelet has sufficient PID available,},NodeCondition{Type:Ready,Status:True,LastHeartbeatTime:2022-11-26 05:11:14 +0000 UTC,LastTransitionTime:2022-11-26 05:05:38 +0000 UTC,Reason:KubeletReady,Message:kubelet is posting ready status. AppArmor enabled,},},Addresses:[]NodeAddress{NodeAddress{Type:InternalIP,Address:10.138.0.2,},NodeAddress{Type:ExternalIP,Address:35.227.179.144,},NodeAddress{Type:InternalDNS,Address:bootstrap-e2e-master.c.k8s-jkns-e2e-gce-ci-reboot.internal,},NodeAddress{Type:Hostname,Address:bootstrap-e2e-master.c.k8s-jkns-e2e-gce-ci-reboot.internal,},},DaemonEndpoints:NodeDaemonEndpoints{KubeletEndpoint:DaemonEndpoint{Port:10250,},},NodeInfo:NodeSystemInfo{MachineID:0fdb3cfe29f66637553465718381a2f8,SystemUUID:0fdb3cfe-29f6-6637-5534-65718381a2f8,BootID:fb804a29-7e61-434d-8044-6944159ce35f,KernelVersion:5.10.123+,OSImage:Container-Optimized OS from Google,ContainerRuntimeVersion:containerd://1.7.0-beta.0-149-gd06318622,KubeletVersion:v1.27.0-alpha.0.50+70617042976dc1,KubeProxyVersion:v1.27.0-alpha.0.50+70617042976dc1,OperatingSystem:linux,Architecture:amd64,},Images:[]ContainerImage{ContainerImage{Names:[registry.k8s.io/kube-apiserver-amd64:v1.27.0-alpha.0.50_70617042976dc1],SizeBytes:135160272,},ContainerImage{Names:[registry.k8s.io/kube-controller-manager-amd64:v1.27.0-alpha.0.50_70617042976dc1],SizeBytes:124990265,},ContainerImage{Names:[registry.k8s.io/etcd@sha256:dd75ec974b0a2a6f6bb47001ba09207976e625db898d1b16735528c009cb171c registry.k8s.io/etcd:3.5.6-0],SizeBytes:102542580,},ContainerImage{Names:[registry.k8s.io/kube-scheduler-amd64:v1.27.0-alpha.0.50_70617042976dc1],SizeBytes:57660216,},ContainerImage{Names:[gke.gcr.io/prometheus-to-sd@sha256:e739643c3939ba0b161425f45a1989eedfc4a3b166db9a7100863296b4c70510 gke.gcr.io/prometheus-to-sd:v0.11.1-gke.1],SizeBytes:48742566,},ContainerImage{Names:[gcr.io/k8s-ingress-image-push/ingress-gce-glbc-amd64@sha256:5db27383add6d9f4ebdf0286409ac31f7f5d273690204b341a4e37998917693b gcr.io/k8s-ingress-image-push/ingress-gce-glbc-amd64:v1.20.1],SizeBytes:36598135,},ContainerImage{Names:[registry.k8s.io/addon-manager/kube-addon-manager@sha256:49cc4e6e4a3745b427ce14b0141476ab339bb65c6bc05033019e046c8727dcb0 registry.k8s.io/addon-manager/kube-addon-manager:v9.1.6],SizeBytes:30464183,},ContainerImage{Names:[registry.k8s.io/kas-network-proxy/proxy-server@sha256:2c111f004bec24888d8cfa2a812a38fb8341350abac67dcd0ac64e709dfe389c registry.k8s.io/kas-network-proxy/proxy-server:v0.0.33],SizeBytes:22020129,},ContainerImage{Names:[registry.k8s.io/metadata-proxy@sha256:e914645f22e946bce5165737e1b244e0a296ad1f0f81a9531adc57af2780978a registry.k8s.io/metadata-proxy:v0.1.12],SizeBytes:5301657,},ContainerImage{Names:[registry.k8s.io/pause@sha256:9001185023633d17a2f98ff69b6ff2615b8ea02a825adffa40422f51dfdcde9d registry.k8s.io/pause:3.8],SizeBytes:311286,},},VolumesInUse:[],VolumesAttached:[]AttachedVolume{},Config:nil,},} Nov 26 05:15:39.335: INFO: Logging kubelet events for node bootstrap-e2e-master Nov 26 05:15:39.381: INFO: Logging pods the kubelet thinks is on node bootstrap-e2e-master Nov 26 05:15:39.445: INFO: etcd-server-events-bootstrap-e2e-master started at 2022-11-26 05:04:35 +0000 UTC (0+1 container statuses recorded) Nov 26 05:15:39.445: INFO: Container etcd-container ready: true, restart count 3 Nov 26 05:15:39.445: INFO: kube-controller-manager-bootstrap-e2e-master started at 2022-11-26 05:04:35 +0000 UTC (0+1 container statuses recorded) Nov 26 05:15:39.445: INFO: Container kube-controller-manager ready: true, restart count 4 Nov 26 05:15:39.445: INFO: kube-scheduler-bootstrap-e2e-master started at 2022-11-26 05:04:35 +0000 UTC (0+1 container statuses recorded) Nov 26 05:15:39.445: INFO: Container kube-scheduler ready: true, restart count 4 Nov 26 05:15:39.445: INFO: kube-addon-manager-bootstrap-e2e-master started at 2022-11-26 05:04:52 +0000 UTC (0+1 container statuses recorded) Nov 26 05:15:39.445: INFO: Container kube-addon-manager ready: true, restart count 1 Nov 26 05:15:39.445: INFO: metadata-proxy-v0.1-8ll2g started at 2022-11-26 05:05:51 +0000 UTC (0+2 container statuses recorded) Nov 26 05:15:39.445: INFO: Container metadata-proxy ready: true, restart count 0 Nov 26 05:15:39.445: INFO: Container prometheus-to-sd-exporter ready: true, restart count 0 Nov 26 05:15:39.445: INFO: etcd-server-bootstrap-e2e-master started at 2022-11-26 05:04:35 +0000 UTC (0+1 container statuses recorded) Nov 26 05:15:39.445: INFO: Container etcd-container ready: true, restart count 2 Nov 26 05:15:39.445: INFO: konnectivity-server-bootstrap-e2e-master started at 2022-11-26 05:04:35 +0000 UTC (0+1 container statuses recorded) Nov 26 05:15:39.445: INFO: Container konnectivity-server-container ready: true, restart count 2 Nov 26 05:15:39.445: INFO: kube-apiserver-bootstrap-e2e-master started at 2022-11-26 05:04:35 +0000 UTC (0+1 container statuses recorded) Nov 26 05:15:39.445: INFO: Container kube-apiserver ready: true, restart count 1 Nov 26 05:15:39.445: INFO: l7-lb-controller-bootstrap-e2e-master started at 2022-11-26 05:04:52 +0000 UTC (0+1 container statuses recorded) Nov 26 05:15:39.445: INFO: Container l7-lb-controller ready: false, restart count 5 Nov 26 05:15:39.618: INFO: Latency metrics for node bootstrap-e2e-master Nov 26 05:15:39.618: INFO: Logging node info for node bootstrap-e2e-minion-group-0975 Nov 26 05:15:39.660: INFO: Node Info: &Node{ObjectMeta:{bootstrap-e2e-minion-group-0975 6a0671d4-0e9d-4ac8-b359-65111262cfad 6516 0 2022-11-26 05:05:22 +0000 UTC <nil> <nil> map[beta.kubernetes.io/arch:amd64 beta.kubernetes.io/instance-type:n1-standard-2 beta.kubernetes.io/os:linux cloud.google.com/metadata-proxy-ready:true failure-domain.beta.kubernetes.io/region:us-west1 failure-domain.beta.kubernetes.io/zone:us-west1-b kubernetes.io/arch:amd64 kubernetes.io/hostname:bootstrap-e2e-minion-group-0975 kubernetes.io/os:linux node.kubernetes.io/instance-type:n1-standard-2 topology.hostpath.csi/node:bootstrap-e2e-minion-group-0975 topology.kubernetes.io/region:us-west1 topology.kubernetes.io/zone:us-west1-b] map[csi.volume.kubernetes.io/nodeid:{"csi-hostpath-multivolume-5830":"bootstrap-e2e-minion-group-0975","csi-hostpath-multivolume-593":"bootstrap-e2e-minion-group-0975","csi-hostpath-multivolume-702":"bootstrap-e2e-minion-group-0975","csi-hostpath-provisioning-7492":"bootstrap-e2e-minion-group-0975"} node.alpha.kubernetes.io/ttl:0 volumes.kubernetes.io/controller-managed-attach-detach:true] [] [] [{kubelet Update v1 2022-11-26 05:05:22 +0000 UTC FieldsV1 {"f:metadata":{"f:annotations":{".":{},"f:volumes.kubernetes.io/controller-managed-attach-detach":{}},"f:labels":{".":{},"f:beta.kubernetes.io/arch":{},"f:beta.kubernetes.io/instance-type":{},"f:beta.kubernetes.io/os":{},"f:cloud.google.com/metadata-proxy-ready":{},"f:failure-domain.beta.kubernetes.io/region":{},"f:failure-domain.beta.kubernetes.io/zone":{},"f:kubernetes.io/arch":{},"f:kubernetes.io/hostname":{},"f:kubernetes.io/os":{},"f:node.kubernetes.io/instance-type":{},"f:topology.kubernetes.io/region":{},"f:topology.kubernetes.io/zone":{}}},"f:spec":{"f:providerID":{}}} } {kube-controller-manager Update v1 2022-11-26 05:05:23 +0000 UTC FieldsV1 {"f:metadata":{"f:annotations":{"f:node.alpha.kubernetes.io/ttl":{}}},"f:spec":{"f:podCIDR":{},"f:podCIDRs":{".":{},"v:\"10.64.2.0/24\"":{}}}} } {kubelet Update v1 2022-11-26 05:13:30 +0000 UTC FieldsV1 {"f:metadata":{"f:annotations":{"f:csi.volume.kubernetes.io/nodeid":{}},"f:labels":{"f:topology.hostpath.csi/node":{}}},"f:status":{"f:conditions":{"k:{\"type\":\"DiskPressure\"}":{"f:lastHeartbeatTime":{}},"k:{\"type\":\"MemoryPressure\"}":{"f:lastHeartbeatTime":{}},"k:{\"type\":\"PIDPressure\"}":{"f:lastHeartbeatTime":{}},"k:{\"type\":\"Ready\"}":{"f:lastHeartbeatTime":{},"f:lastTransitionTime":{},"f:message":{},"f:reason":{},"f:status":{}}},"f:images":{},"f:volumesInUse":{}}} status} {kube-controller-manager Update v1 2022-11-26 05:13:31 +0000 UTC FieldsV1 {"f:status":{"f:conditions":{"k:{\"type\":\"NetworkUnavailable\"}":{"f:lastHeartbeatTime":{},"f:lastTransitionTime":{},"f:message":{},"f:reason":{},"f:status":{}}},"f:volumesAttached":{}}} status} {node-problem-detector Update v1 2022-11-26 05:15:29 +0000 UTC FieldsV1 {"f:status":{"f:conditions":{"k:{\"type\":\"CorruptDockerOverlay2\"}":{".":{},"f:lastHeartbeatTime":{},"f:lastTransitionTime":{},"f:message":{},"f:reason":{},"f:status":{},"f:type":{}},"k:{\"type\":\"FrequentContainerdRestart\"}":{".":{},"f:lastHeartbeatTime":{},"f:lastTransitionTime":{},"f:message":{},"f:reason":{},"f:status":{},"f:type":{}},"k:{\"type\":\"FrequentDockerRestart\"}":{".":{},"f:lastHeartbeatTime":{},"f:lastTransitionTime":{},"f:message":{},"f:reason":{},"f:status":{},"f:type":{}},"k:{\"type\":\"FrequentKubeletRestart\"}":{".":{},"f:lastHeartbeatTime":{},"f:lastTransitionTime":{},"f:message":{},"f:reason":{},"f:status":{},"f:type":{}},"k:{\"type\":\"FrequentUnregisterNetDevice\"}":{".":{},"f:lastHeartbeatTime":{},"f:lastTransitionTime":{},"f:message":{},"f:reason":{},"f:status":{},"f:type":{}},"k:{\"type\":\"KernelDeadlock\"}":{".":{},"f:lastHeartbeatTime":{},"f:lastTransitionTime":{},"f:message":{},"f:reason":{},"f:status":{},"f:type":{}},"k:{\"type\":\"ReadonlyFilesystem\"}":{".":{},"f:lastHeartbeatTime":{},"f:lastTransitionTime":{},"f:message":{},"f:reason":{},"f:status":{},"f:type":{}}}}} status}]},Spec:NodeSpec{PodCIDR:10.64.2.0/24,DoNotUseExternalID:,ProviderID:gce://k8s-jkns-e2e-gce-ci-reboot/us-west1-b/bootstrap-e2e-minion-group-0975,Unschedulable:false,Taints:[]Taint{},ConfigSource:nil,PodCIDRs:[10.64.2.0/24],},Status:NodeStatus{Capacity:ResourceList{attachable-volumes-gce-pd: {{127 0} {<nil>} 127 DecimalSI},cpu: {{2 0} {<nil>} 2 DecimalSI},ephemeral-storage: {{101203873792 0} {<nil>} 98831908Ki BinarySI},hugepages-1Gi: {{0 0} {<nil>} 0 DecimalSI},hugepages-2Mi: {{0 0} {<nil>} 0 DecimalSI},memory: {{7815430144 0} {<nil>} BinarySI},pods: {{110 0} {<nil>} 110 DecimalSI},},Allocatable:ResourceList{attachable-volumes-gce-pd: {{127 0} {<nil>} 127 DecimalSI},cpu: {{2 0} {<nil>} 2 DecimalSI},ephemeral-storage: {{91083486262 0} {<nil>} 91083486262 DecimalSI},hugepages-1Gi: {{0 0} {<nil>} 0 DecimalSI},hugepages-2Mi: {{0 0} {<nil>} 0 DecimalSI},memory: {{7553286144 0} {<nil>} BinarySI},pods: {{110 0} {<nil>} 110 DecimalSI},},Phase:,Conditions:[]NodeCondition{NodeCondition{Type:FrequentUnregisterNetDevice,Status:False,LastHeartbeatTime:2022-11-26 05:15:29 +0000 UTC,LastTransitionTime:2022-11-26 05:05:27 +0000 UTC,Reason:NoFrequentUnregisterNetDevice,Message:node is functioning properly,},NodeCondition{Type:FrequentKubeletRestart,Status:False,LastHeartbeatTime:2022-11-26 05:15:29 +0000 UTC,LastTransitionTime:2022-11-26 05:05:27 +0000 UTC,Reason:NoFrequentKubeletRestart,Message:kubelet is functioning properly,},NodeCondition{Type:FrequentDockerRestart,Status:False,LastHeartbeatTime:2022-11-26 05:15:29 +0000 UTC,LastTransitionTime:2022-11-26 05:05:27 +0000 UTC,Reason:NoFrequentDockerRestart,Message:docker is functioning properly,},NodeCondition{Type:FrequentContainerdRestart,Status:False,LastHeartbeatTime:2022-11-26 05:15:29 +0000 UTC,LastTransitionTime:2022-11-26 05:05:27 +0000 UTC,Reason:NoFrequentContainerdRestart,Message:containerd is functioning properly,},NodeCondition{Type:KernelDeadlock,Status:False,LastHeartbeatTime:2022-11-26 05:15:29 +0000 UTC,LastTransitionTime:2022-11-26 05:05:27 +0000 UTC,Reason:KernelHasNoDeadlock,Message:kernel has no deadlock,},NodeCondition{Type:ReadonlyFilesystem,Status:False,LastHeartbeatTime:2022-11-26 05:15:29 +0000 UTC,LastTransitionTime:2022-11-26 05:05:27 +0000 UTC,Reason:FilesystemIsNotReadOnly,Message:Filesystem is not read-only,},NodeCondition{Type:CorruptDockerOverlay2,Status:False,LastHeartbeatTime:2022-11-26 05:15:29 +0000 UTC,LastTransitionTime:2022-11-26 05:05:27 +0000 UTC,Reason:NoCorruptDockerOverlay2,Message:docker overlay2 is functioning properly,},NodeCondition{Type:NetworkUnavailable,Status:False,LastHeartbeatTime:2022-11-26 05:05:39 +0000 UTC,LastTransitionTime:2022-11-26 05:05:39 +0000 UTC,Reason:RouteCreated,Message:RouteController created a route,},NodeCondition{Type:MemoryPressure,Status:False,LastHeartbeatTime:2022-11-26 05:13:30 +0000 UTC,LastTransitionTime:2022-11-26 05:05:22 +0000 UTC,Reason:KubeletHasSufficientMemory,Message:kubelet has sufficient memory available,},NodeCondition{Type:DiskPressure,Status:False,LastHeartbeatTime:2022-11-26 05:13:30 +0000 UTC,LastTransitionTime:2022-11-26 05:05:22 +0000 UTC,Reason:KubeletHasNoDiskPressure,Message:kubelet has no disk pressure,},NodeCondition{Type:PIDPressure,Status:False,LastHeartbeatTime:2022-11-26 05:13:30 +0000 UTC,LastTransitionTime:2022-11-26 05:05:22 +0000 UTC,Reason:KubeletHasSufficientPID,Message:kubelet has sufficient PID available,},NodeCondition{Type:Ready,Status:True,LastHeartbeatTime:2022-11-26 05:13:30 +0000 UTC,LastTransitionTime:2022-11-26 05:05:23 +0000 UTC,Reason:KubeletReady,Message:kubelet is posting ready status. AppArmor enabled,},},Addresses:[]NodeAddress{NodeAddress{Type:InternalIP,Address:10.138.0.5,},NodeAddress{Type:ExternalIP,Address:35.247.104.236,},NodeAddress{Type:InternalDNS,Address:bootstrap-e2e-minion-group-0975.c.k8s-jkns-e2e-gce-ci-reboot.internal,},NodeAddress{Type:Hostname,Address:bootstrap-e2e-minion-group-0975.c.k8s-jkns-e2e-gce-ci-reboot.internal,},},DaemonEndpoints:NodeDaemonEndpoints{KubeletEndpoint:DaemonEndpoint{Port:10250,},},NodeInfo:NodeSystemInfo{MachineID:b38385f7976a474936da82f5f565760a,SystemUUID:b38385f7-976a-4749-36da-82f5f565760a,BootID:0aa74273-9b8d-48ac-bad0-0a1805f010ea,KernelVersion:5.10.123+,OSImage:Container-Optimized OS from Google,ContainerRuntimeVersion:containerd://1.7.0-beta.0-149-gd06318622,KubeletVersion:v1.27.0-alpha.0.50+70617042976dc1,KubeProxyVersion:v1.27.0-alpha.0.50+70617042976dc1,OperatingSystem:linux,Architecture:amd64,},Images:[]ContainerImage{ContainerImage{Names:[registry.k8s.io/e2e-test-images/jessie-dnsutils@sha256:24aaf2626d6b27864c29de2097e8bbb840b3a414271bf7c8995e431e47d8408e registry.k8s.io/e2e-test-images/jessie-dnsutils:1.7],SizeBytes:112030336,},ContainerImage{Names:[registry.k8s.io/sig-storage/nfs-provisioner@sha256:e943bb77c7df05ebdc8c7888b2db289b13bf9f012d6a3a5a74f14d4d5743d439 registry.k8s.io/sig-storage/nfs-provisioner:v3.0.1],SizeBytes:90632047,},ContainerImage{Names:[registry.k8s.io/kube-proxy-amd64:v1.27.0-alpha.0.50_70617042976dc1],SizeBytes:67201736,},ContainerImage{Names:[registry.k8s.io/e2e-test-images/agnhost@sha256:16bbf38c463a4223d8cfe4da12bc61010b082a79b4bb003e2d3ba3ece5dd5f9e registry.k8s.io/e2e-test-images/agnhost:2.43],SizeBytes:51706353,},ContainerImage{Names:[gke.gcr.io/prometheus-to-sd@sha256:e739643c3939ba0b161425f45a1989eedfc4a3b166db9a7100863296b4c70510 gke.gcr.io/prometheus-to-sd:v0.11.1-gke.1],SizeBytes:48742566,},ContainerImage{Names:[registry.k8s.io/e2e-test-images/httpd@sha256:148b022f5c5da426fc2f3c14b5c0867e58ef05961510c84749ac1fddcb0fef22 registry.k8s.io/e2e-test-images/httpd:2.4.38-4],SizeBytes:40764257,},ContainerImage{Names:[registry.k8s.io/metrics-server/metrics-server@sha256:6385aec64bb97040a5e692947107b81e178555c7a5b71caa90d733e4130efc10 registry.k8s.io/metrics-server/metrics-server:v0.5.2],SizeBytes:26023008,},ContainerImage{Names:[registry.k8s.io/sig-storage/csi-provisioner@sha256:ee3b525d5b89db99da3b8eb521d9cd90cb6e9ef0fbb651e98bb37be78d36b5b8 registry.k8s.io/sig-storage/csi-provisioner:v3.3.0],SizeBytes:25491225,},ContainerImage{Names:[registry.k8s.io/sig-storage/csi-resizer@sha256:425d8f1b769398127767b06ed97ce62578a3179bcb99809ce93a1649e025ffe7 registry.k8s.io/sig-storage/csi-resizer:v1.6.0],SizeBytes:24148884,},ContainerImage{Names:[registry.k8s.io/sig-storage/csi-snapshotter@sha256:291334908ddf71a4661fd7f6d9d97274de8a5378a2b6fdfeb2ce73414a34f82f registry.k8s.io/sig-storage/csi-snapshotter:v6.1.0],SizeBytes:23881995,},ContainerImage{Names:[registry.k8s.io/sig-storage/csi-attacher@sha256:9a685020911e2725ad019dbce6e4a5ab93d51e3d4557f115e64343345e05781b registry.k8s.io/sig-storage/csi-attacher:v4.0.0],SizeBytes:23847201,},ContainerImage{Names:[registry.k8s.io/sig-storage/hostpathplugin@sha256:92257881c1d6493cf18299a24af42330f891166560047902b8d431fb66b01af5 registry.k8s.io/sig-storage/hostpathplugin:v1.9.0],SizeBytes:18758628,},ContainerImage{Names:[registry.k8s.io/autoscaling/addon-resizer@sha256:43f129b81d28f0fdd54de6d8e7eacd5728030782e03db16087fc241ad747d3d6 registry.k8s.io/autoscaling/addon-resizer:1.8.14],SizeBytes:10153852,},ContainerImage{Names:[registry.k8s.io/sig-storage/csi-node-driver-registrar@sha256:0103eee7c35e3e0b5cd8cdca9850dc71c793cdeb6669d8be7a89440da2d06ae4 registry.k8s.io/sig-storage/csi-node-driver-registrar:v2.5.1],SizeBytes:9133109,},ContainerImage{Names:[registry.k8s.io/sig-storage/livenessprobe@sha256:933940f13b3ea0abc62e656c1aa5c5b47c04b15d71250413a6b821bd0c58b94e registry.k8s.io/sig-storage/livenessprobe:v2.7.0],SizeBytes:8688564,},ContainerImage{Names:[registry.k8s.io/kas-network-proxy/proxy-agent@sha256:48f2a4ec3e10553a81b8dd1c6fa5fe4bcc9617f78e71c1ca89c6921335e2d7da registry.k8s.io/kas-network-proxy/proxy-agent:v0.0.33],SizeBytes:8512162,},ContainerImage{Names:[registry.k8s.io/metadata-proxy@sha256:e914645f22e946bce5165737e1b244e0a296ad1f0f81a9531adc57af2780978a registry.k8s.io/metadata-proxy:v0.1.12],SizeBytes:5301657,},ContainerImage{Names:[registry.k8s.io/e2e-test-images/busybox@sha256:2e0f836850e09b8b7cc937681d6194537a09fbd5f6b9e08f4d646a85128e8937 registry.k8s.io/e2e-test-images/busybox:1.29-4],SizeBytes:731990,},ContainerImage{Names:[registry.k8s.io/pause@sha256:9001185023633d17a2f98ff69b6ff2615b8ea02a825adffa40422f51dfdcde9d registry.k8s.io/pause:3.8],SizeBytes:311286,},},VolumesInUse:[kubernetes.io/csi/csi-hostpath-multivolume-702^02fad0b0-6d49-11ed-9d4c-26ec58c28662],VolumesAttached:[]AttachedVolume{AttachedVolume{Name:kubernetes.io/csi/csi-hostpath-multivolume-702^02fad0b0-6d49-11ed-9d4c-26ec58c28662,DevicePath:,},},Config:nil,},} Nov 26 05:15:39.660: INFO: Logging kubelet events for node bootstrap-e2e-minion-group-0975 Nov 26 05:15:39.704: INFO: Logging pods the kubelet thinks is on node bootstrap-e2e-minion-group-0975 Nov 26 05:15:39.796: INFO: metrics-server-v0.5.2-867b8754b9-27xcg started at 2022-11-26 05:05:51 +0000 UTC (0+2 container statuses recorded) Nov 26 05:15:39.796: INFO: Container metrics-server ready: false, restart count 5 Nov 26 05:15:39.796: INFO: Container metrics-server-nanny ready: false, restart count 6 Nov 26 05:15:39.796: INFO: csi-hostpathplugin-0 started at 2022-11-26 05:09:53 +0000 UTC (0+7 container statuses recorded) Nov 26 05:15:39.796: INFO: Container csi-attacher ready: true, restart count 2 Nov 26 05:15:39.796: INFO: Container csi-provisioner ready: true, restart count 2 Nov 26 05:15:39.796: INFO: Container csi-resizer ready: true, restart count 2 Nov 26 05:15:39.796: INFO: Container csi-snapshotter ready: true, restart count 2 Nov 26 05:15:39.796: INFO: Container hostpath ready: true, restart count 2 Nov 26 05:15:39.796: INFO: Container liveness-probe ready: true, restart count 2 Nov 26 05:15:39.796: INFO: Container node-driver-registrar ready: true, restart count 2 Nov 26 05:15:39.796: INFO: affinity-lb-esipp-transition-8sg8w started at 2022-11-26 05:10:44 +0000 UTC (0+1 container statuses recorded) Nov 26 05:15:39.796: INFO: Container affinity-lb-esipp-transition ready: true, restart count 1 Nov 26 05:15:39.796: INFO: external-local-update-5xshb started at 2022-11-26 05:11:23 +0000 UTC (0+1 container statuses recorded) Nov 26 05:15:39.796: INFO: Container netexec ready: false, restart count 4 Nov 26 05:15:39.796: INFO: host-test-container-pod started at 2022-11-26 05:12:10 +0000 UTC (0+1 container statuses recorded) Nov 26 05:15:39.796: INFO: Container agnhost-container ready: true, restart count 2 Nov 26 05:15:39.796: INFO: pod-db90ef24-4c1c-4a0d-81ed-e023c5963392 started at 2022-11-26 05:13:08 +0000 UTC (0+1 container statuses recorded) Nov 26 05:15:39.796: INFO: Container write-pod ready: false, restart count 0 Nov 26 05:15:39.796: INFO: metadata-proxy-v0.1-mznrc started at 2022-11-26 05:05:24 +0000 UTC (0+2 container statuses recorded) Nov 26 05:15:39.796: INFO: Container metadata-proxy ready: true, restart count 0 Nov 26 05:15:39.796: INFO: Container prometheus-to-sd-exporter ready: true, restart count 0 Nov 26 05:15:39.796: INFO: konnectivity-agent-wfc29 started at 2022-11-26 05:05:39 +0000 UTC (0+1 container statuses recorded) Nov 26 05:15:39.796: INFO: Container konnectivity-agent ready: true, restart count 6 Nov 26 05:15:39.796: INFO: netserver-0 started at 2022-11-26 05:11:06 +0000 UTC (0+1 container statuses recorded) Nov 26 05:15:39.796: INFO: Container webserver ready: true, restart count 3 Nov 26 05:15:39.796: INFO: csi-hostpathplugin-0 started at 2022-11-26 05:12:10 +0000 UTC (0+7 container statuses recorded) Nov 26 05:15:39.796: INFO: Container csi-attacher ready: true, restart count 1 Nov 26 05:15:39.796: INFO: Container csi-provisioner ready: true, restart count 1 Nov 26 05:15:39.796: INFO: Container csi-resizer ready: true, restart count 1 Nov 26 05:15:39.796: INFO: Container csi-snapshotter ready: true, restart count 1 Nov 26 05:15:39.796: INFO: Container hostpath ready: true, restart count 1 Nov 26 05:15:39.796: INFO: Container liveness-probe ready: true, restart count 1 Nov 26 05:15:39.796: INFO: Container node-driver-registrar ready: true, restart count 1 Nov 26 05:15:39.796: INFO: pod-secrets-ae6e2110-abaa-4433-ab33-c417114f813f started at 2022-11-26 05:10:35 +0000 UTC (0+1 container statuses recorded) Nov 26 05:15:39.796: INFO: Container creates-volume-test ready: false, restart count 0 Nov 26 05:15:39.796: INFO: csi-hostpathplugin-0 started at 2022-11-26 05:13:04 +0000 UTC (0+7 container statuses recorded) Nov 26 05:15:39.796: INFO: Container csi-attacher ready: true, restart count 0 Nov 26 05:15:39.796: INFO: Container csi-provisioner ready: true, restart count 0 Nov 26 05:15:39.796: INFO: Container csi-resizer ready: true, restart count 0 Nov 26 05:15:39.796: INFO: Container csi-snapshotter ready: true, restart count 0 Nov 26 05:15:39.796: INFO: Container hostpath ready: true, restart count 0 Nov 26 05:15:39.796: INFO: Container liveness-probe ready: true, restart count 0 Nov 26 05:15:39.796: INFO: Container node-driver-registrar ready: true, restart count 0 Nov 26 05:15:39.796: INFO: external-provisioner-ntc44 started at 2022-11-26 05:11:08 +0000 UTC (0+1 container statuses recorded) Nov 26 05:15:39.796: INFO: Container nfs-provisioner ready: true, restart count 2 Nov 26 05:15:39.796: INFO: kube-proxy-bootstrap-e2e-minion-group-0975 started at 2022-11-26 05:05:23 +0000 UTC (0+1 container statuses recorded) Nov 26 05:15:39.796: INFO: Container kube-proxy ready: false, restart count 4 Nov 26 05:15:39.796: INFO: csi-hostpathplugin-0 started at 2022-11-26 05:13:03 +0000 UTC (0+7 container statuses recorded) Nov 26 05:15:39.796: INFO: Container csi-attacher ready: true, restart count 1 Nov 26 05:15:39.796: INFO: Container csi-provisioner ready: true, restart count 1 Nov 26 05:15:39.796: INFO: Container csi-resizer ready: true, restart count 1 Nov 26 05:15:39.796: INFO: Container csi-snapshotter ready: true, restart count 1 Nov 26 05:15:39.796: INFO: Container hostpath ready: true, restart count 1 Nov 26 05:15:39.796: INFO: Container liveness-probe ready: true, restart count 1 Nov 26 05:15:39.796: INFO: Container node-driver-registrar ready: true, restart count 1 Nov 26 05:15:39.796: INFO: netserver-0 started at 2022-11-26 05:11:29 +0000 UTC (0+1 container statuses recorded) Nov 26 05:15:39.796: INFO: Container webserver ready: true, restart count 3 Nov 26 05:15:39.796: INFO: netserver-0 started at 2022-11-26 05:09:51 +0000 UTC (0+1 container statuses recorded) Nov 26 05:15:39.796: INFO: Container webserver ready: false, restart count 4 Nov 26 05:15:39.796: INFO: inclusterclient started at 2022-11-26 05:12:36 +0000 UTC (0+1 container statuses recorded) Nov 26 05:15:39.796: INFO: Container inclusterclient ready: false, restart count 0 Nov 26 05:15:40.060: INFO: Latency metrics for node bootstrap-e2e-minion-group-0975 Nov 26 05:15:40.061: INFO: Logging node info for node bootstrap-e2e-minion-group-rwsn Nov 26 05:15:40.102: INFO: Node Info: &Node{ObjectMeta:{bootstrap-e2e-minion-group-rwsn adedf225-b462-4305-a60c-9042d31ca1d2 6492 0 2022-11-26 05:05:15 +0000 UTC <nil> <nil> map[beta.kubernetes.io/arch:amd64 beta.kubernetes.io/instance-type:n1-standard-2 beta.kubernetes.io/os:linux cloud.google.com/metadata-proxy-ready:true failure-domain.beta.kubernetes.io/region:us-west1 failure-domain.beta.kubernetes.io/zone:us-west1-b kubernetes.io/arch:amd64 kubernetes.io/hostname:bootstrap-e2e-minion-group-rwsn kubernetes.io/os:linux node.kubernetes.io/instance-type:n1-standard-2 topology.hostpath.csi/node:bootstrap-e2e-minion-group-rwsn topology.kubernetes.io/region:us-west1 topology.kubernetes.io/zone:us-west1-b] map[csi.volume.kubernetes.io/nodeid:{"csi-mock-csi-mock-volumes-8906":"csi-mock-csi-mock-volumes-8906"} node.alpha.kubernetes.io/ttl:0 volumes.kubernetes.io/controller-managed-attach-detach:true] [] [] [{kubelet Update v1 2022-11-26 05:05:15 +0000 UTC FieldsV1 {"f:metadata":{"f:annotations":{".":{},"f:volumes.kubernetes.io/controller-managed-attach-detach":{}},"f:labels":{".":{},"f:beta.kubernetes.io/arch":{},"f:beta.kubernetes.io/instance-type":{},"f:beta.kubernetes.io/os":{},"f:cloud.google.com/metadata-proxy-ready":{},"f:failure-domain.beta.kubernetes.io/region":{},"f:failure-domain.beta.kubernetes.io/zone":{},"f:kubernetes.io/arch":{},"f:kubernetes.io/hostname":{},"f:kubernetes.io/os":{},"f:node.kubernetes.io/instance-type":{},"f:topology.kubernetes.io/region":{},"f:topology.kubernetes.io/zone":{}}},"f:spec":{"f:providerID":{}}} } {kube-controller-manager Update v1 2022-11-26 05:05:17 +0000 UTC FieldsV1 {"f:metadata":{"f:annotations":{"f:node.alpha.kubernetes.io/ttl":{}}},"f:spec":{"f:podCIDR":{},"f:podCIDRs":{".":{},"v:\"10.64.0.0/24\"":{}}}} } {kube-controller-manager Update v1 2022-11-26 05:08:09 +0000 UTC FieldsV1 {"f:status":{"f:conditions":{"k:{\"type\":\"NetworkUnavailable\"}":{"f:lastHeartbeatTime":{},"f:lastTransitionTime":{},"f:message":{},"f:reason":{},"f:status":{}}},"f:volumesAttached":{}}} status} {kubelet Update v1 2022-11-26 05:14:49 +0000 UTC FieldsV1 {"f:metadata":{"f:annotations":{"f:csi.volume.kubernetes.io/nodeid":{}},"f:labels":{"f:topology.hostpath.csi/node":{}}},"f:status":{"f:conditions":{"k:{\"type\":\"DiskPressure\"}":{"f:lastHeartbeatTime":{}},"k:{\"type\":\"MemoryPressure\"}":{"f:lastHeartbeatTime":{}},"k:{\"type\":\"PIDPressure\"}":{"f:lastHeartbeatTime":{}},"k:{\"type\":\"Ready\"}":{"f:lastHeartbeatTime":{},"f:lastTransitionTime":{},"f:message":{},"f:reason":{},"f:status":{}}},"f:images":{},"f:volumesInUse":{}}} status} {node-problem-detector Update v1 2022-11-26 05:15:20 +0000 UTC FieldsV1 {"f:status":{"f:conditions":{"k:{\"type\":\"CorruptDockerOverlay2\"}":{".":{},"f:lastHeartbeatTime":{},"f:lastTransitionTime":{},"f:message":{},"f:reason":{},"f:status":{},"f:type":{}},"k:{\"type\":\"FrequentContainerdRestart\"}":{".":{},"f:lastHeartbeatTime":{},"f:lastTransitionTime":{},"f:message":{},"f:reason":{},"f:status":{},"f:type":{}},"k:{\"type\":\"FrequentDockerRestart\"}":{".":{},"f:lastHeartbeatTime":{},"f:lastTransitionTime":{},"f:message":{},"f:reason":{},"f:status":{},"f:type":{}},"k:{\"type\":\"FrequentKubeletRestart\"}":{".":{},"f:lastHeartbeatTime":{},"f:lastTransitionTime":{},"f:message":{},"f:reason":{},"f:status":{},"f:type":{}},"k:{\"type\":\"FrequentUnregisterNetDevice\"}":{".":{},"f:lastHeartbeatTime":{},"f:lastTransitionTime":{},"f:message":{},"f:reason":{},"f:status":{},"f:type":{}},"k:{\"type\":\"KernelDeadlock\"}":{".":{},"f:lastHeartbeatTime":{},"f:lastTransitionTime":{},"f:message":{},"f:reason":{},"f:status":{},"f:type":{}},"k:{\"type\":\"ReadonlyFilesystem\"}":{".":{},"f:lastHeartbeatTime":{},"f:lastTransitionTime":{},"f:message":{},"f:reason":{},"f:status":{},"f:type":{}}}}} status}]},Spec:NodeSpec{PodCIDR:10.64.0.0/24,DoNotUseExternalID:,ProviderID:gce://k8s-jkns-e2e-gce-ci-reboot/us-west1-b/bootstrap-e2e-minion-group-rwsn,Unschedulable:false,Taints:[]Taint{},ConfigSource:nil,PodCIDRs:[10.64.0.0/24],},Status:NodeStatus{Capacity:ResourceList{attachable-volumes-gce-pd: {{127 0} {<nil>} 127 DecimalSI},cpu: {{2 0} {<nil>} 2 DecimalSI},ephemeral-storage: {{101203873792 0} {<nil>} 98831908Ki BinarySI},hugepages-1Gi: {{0 0} {<nil>} 0 DecimalSI},hugepages-2Mi: {{0 0} {<nil>} 0 DecimalSI},memory: {{7815438336 0} {<nil>} BinarySI},pods: {{110 0} {<nil>} 110 DecimalSI},},Allocatable:ResourceList{attachable-volumes-gce-pd: {{127 0} {<nil>} 127 DecimalSI},cpu: {{2 0} {<nil>} 2 DecimalSI},ephemeral-storage: {{91083486262 0} {<nil>} 91083486262 DecimalSI},hugepages-1Gi: {{0 0} {<nil>} 0 DecimalSI},hugepages-2Mi: {{0 0} {<nil>} 0 DecimalSI},memory: {{7553294336 0} {<nil>} BinarySI},pods: {{110 0} {<nil>} 110 DecimalSI},},Phase:,Conditions:[]NodeCondition{NodeCondition{Type:KernelDeadlock,Status:False,LastHeartbeatTime:2022-11-26 05:15:20 +0000 UTC,LastTransitionTime:2022-11-26 05:05:18 +0000 UTC,Reason:KernelHasNoDeadlock,Message:kernel has no deadlock,},NodeCondition{Type:ReadonlyFilesystem,Status:False,LastHeartbeatTime:2022-11-26 05:15:20 +0000 UTC,LastTransitionTime:2022-11-26 05:05:18 +0000 UTC,Reason:FilesystemIsNotReadOnly,Message:Filesystem is not read-only,},NodeCondition{Type:CorruptDockerOverlay2,Status:False,LastHeartbeatTime:2022-11-26 05:15:20 +0000 UTC,LastTransitionTime:2022-11-26 05:05:18 +0000 UTC,Reason:NoCorruptDockerOverlay2,Message:docker overlay2 is functioning properly,},NodeCondition{Type:FrequentUnregisterNetDevice,Status:False,LastHeartbeatTime:2022-11-26 05:15:20 +0000 UTC,LastTransitionTime:2022-11-26 05:05:18 +0000 UTC,Reason:NoFrequentUnregisterNetDevice,Message:node is functioning properly,},NodeCondition{Type:FrequentKubeletRestart,Status:False,LastHeartbeatTime:2022-11-26 05:15:20 +0000 UTC,LastTransitionTime:2022-11-26 05:05:18 +0000 UTC,Reason:NoFrequentKubeletRestart,Message:kubelet is functioning properly,},NodeCondition{Type:FrequentDockerRestart,Status:False,LastHeartbeatTime:2022-11-26 05:15:20 +0000 UTC,LastTransitionTime:2022-11-26 05:05:18 +0000 UTC,Reason:NoFrequentDockerRestart,Message:docker is functioning properly,},NodeCondition{Type:FrequentContainerdRestart,Status:False,LastHeartbeatTime:2022-11-26 05:15:20 +0000 UTC,LastTransitionTime:2022-11-26 05:05:18 +0000 UTC,Reason:NoFrequentContainerdRestart,Message:containerd is functioning properly,},NodeCondition{Type:NetworkUnavailable,Status:False,LastHeartbeatTime:2022-11-26 05:05:25 +0000 UTC,LastTransitionTime:2022-11-26 05:05:25 +0000 UTC,Reason:RouteCreated,Message:RouteController created a route,},NodeCondition{Type:MemoryPressure,Status:False,LastHeartbeatTime:2022-11-26 05:14:43 +0000 UTC,LastTransitionTime:2022-11-26 05:05:15 +0000 UTC,Reason:KubeletHasSufficientMemory,Message:kubelet has sufficient memory available,},NodeCondition{Type:DiskPressure,Status:False,LastHeartbeatTime:2022-11-26 05:14:43 +0000 UTC,LastTransitionTime:2022-11-26 05:05:15 +0000 UTC,Reason:KubeletHasNoDiskPressure,Message:kubelet has no disk pressure,},NodeCondition{Type:PIDPressure,Status:False,LastHeartbeatTime:2022-11-26 05:14:43 +0000 UTC,LastTransitionTime:2022-11-26 05:05:15 +0000 UTC,Reason:KubeletHasSufficientPID,Message:kubelet has sufficient PID available,},NodeCondition{Type:Ready,Status:True,LastHeartbeatTime:2022-11-26 05:14:43 +0000 UTC,LastTransitionTime:2022-11-26 05:05:17 +0000 UTC,Reason:KubeletReady,Message:kubelet is posting ready status. AppArmor enabled,},},Addresses:[]NodeAddress{NodeAddress{Type:InternalIP,Address:10.138.0.4,},NodeAddress{Type:ExternalIP,Address:35.230.27.45,},NodeAddress{Type:InternalDNS,Address:bootstrap-e2e-minion-group-rwsn.c.k8s-jkns-e2e-gce-ci-reboot.internal,},NodeAddress{Type:Hostname,Address:bootstrap-e2e-minion-group-rwsn.c.k8s-jkns-e2e-gce-ci-reboot.internal,},},DaemonEndpoints:NodeDaemonEndpoints{KubeletEndpoint:DaemonEndpoint{Port:10250,},},NodeInfo:NodeSystemInfo{MachineID:69e22ce1281686d6cde1b3211b34b8af,SystemUUID:69e22ce1-2816-86d6-cde1-b3211b34b8af,BootID:79e9e1fd-2a39-42ec-b72b-72836d213862,KernelVersion:5.10.123+,OSImage:Container-Optimized OS from Google,ContainerRuntimeVersion:containerd://1.7.0-beta.0-149-gd06318622,KubeletVersion:v1.27.0-alpha.0.50+70617042976dc1,KubeProxyVersion:v1.27.0-alpha.0.50+70617042976dc1,OperatingSystem:linux,Architecture:amd64,},Images:[]ContainerImage{ContainerImage{Names:[registry.k8s.io/e2e-test-images/volume/nfs@sha256:3bda73f2428522b0e342af80a0b9679e8594c2126f2b3cca39ed787589741b9e registry.k8s.io/e2e-test-images/volume/nfs:1.3],SizeBytes:95836203,},ContainerImage{Names:[registry.k8s.io/sig-storage/nfs-provisioner@sha256:e943bb77c7df05ebdc8c7888b2db289b13bf9f012d6a3a5a74f14d4d5743d439 registry.k8s.io/sig-storage/nfs-provisioner:v3.0.1],SizeBytes:90632047,},ContainerImage{Names:[registry.k8s.io/kube-proxy-amd64:v1.27.0-alpha.0.50_70617042976dc1],SizeBytes:67201736,},ContainerImage{Names:[registry.k8s.io/e2e-test-images/agnhost@sha256:16bbf38c463a4223d8cfe4da12bc61010b082a79b4bb003e2d3ba3ece5dd5f9e registry.k8s.io/e2e-test-images/agnhost:2.43],SizeBytes:51706353,},ContainerImage{Names:[gke.gcr.io/prometheus-to-sd@sha256:e739643c3939ba0b161425f45a1989eedfc4a3b166db9a7100863296b4c70510 gke.gcr.io/prometheus-to-sd:v0.11.1-gke.1],SizeBytes:48742566,},ContainerImage{Names:[registry.k8s.io/metrics-server/metrics-server@sha256:6385aec64bb97040a5e692947107b81e178555c7a5b71caa90d733e4130efc10 registry.k8s.io/metrics-server/metrics-server:v0.5.2],SizeBytes:26023008,},ContainerImage{Names:[registry.k8s.io/sig-storage/csi-provisioner@sha256:ee3b525d5b89db99da3b8eb521d9cd90cb6e9ef0fbb651e98bb37be78d36b5b8 registry.k8s.io/sig-storage/csi-provisioner:v3.3.0],SizeBytes:25491225,},ContainerImage{Names:[registry.k8s.io/sig-storage/csi-resizer@sha256:425d8f1b769398127767b06ed97ce62578a3179bcb99809ce93a1649e025ffe7 registry.k8s.io/sig-storage/csi-resizer:v1.6.0],SizeBytes:24148884,},ContainerImage{Names:[registry.k8s.io/sig-storage/csi-snapshotter@sha256:291334908ddf71a4661fd7f6d9d97274de8a5378a2b6fdfeb2ce73414a34f82f registry.k8s.io/sig-storage/csi-snapshotter:v6.1.0],SizeBytes:23881995,},ContainerImage{Names:[registry.k8s.io/sig-storage/csi-attacher@sha256:9a685020911e2725ad019dbce6e4a5ab93d51e3d4557f115e64343345e05781b registry.k8s.io/sig-storage/csi-attacher:v4.0.0],SizeBytes:23847201,},ContainerImage{Names:[registry.k8s.io/sig-storage/snapshot-controller@sha256:823c75d0c45d1427f6d850070956d9ca657140a7bbf828381541d1d808475280 registry.k8s.io/sig-storage/snapshot-controller:v6.1.0],SizeBytes:22620891,},ContainerImage{Names:[registry.k8s.io/sig-storage/hostpathplugin@sha256:92257881c1d6493cf18299a24af42330f891166560047902b8d431fb66b01af5 registry.k8s.io/sig-storage/hostpathplugin:v1.9.0],SizeBytes:18758628,},ContainerImage{Names:[registry.k8s.io/cpa/cluster-proportional-autoscaler@sha256:fd636b33485c7826fb20ef0688a83ee0910317dbb6c0c6f3ad14661c1db25def registry.k8s.io/cpa/cluster-proportional-autoscaler:1.8.4],SizeBytes:15209393,},ContainerImage{Names:[registry.k8s.io/coredns/coredns@sha256:8e352a029d304ca7431c6507b56800636c321cb52289686a581ab70aaa8a2e2a registry.k8s.io/coredns/coredns:v1.9.3],SizeBytes:14837849,},ContainerImage{Names:[registry.k8s.io/autoscaling/addon-resizer@sha256:43f129b81d28f0fdd54de6d8e7eacd5728030782e03db16087fc241ad747d3d6 registry.k8s.io/autoscaling/addon-resizer:1.8.14],SizeBytes:10153852,},ContainerImage{Names:[registry.k8s.io/sig-storage/csi-node-driver-registrar@sha256:0103eee7c35e3e0b5cd8cdca9850dc71c793cdeb6669d8be7a89440da2d06ae4 registry.k8s.io/sig-storage/csi-node-driver-registrar:v2.5.1],SizeBytes:9133109,},ContainerImage{Names:[registry.k8s.io/sig-storage/livenessprobe@sha256:933940f13b3ea0abc62e656c1aa5c5b47c04b15d71250413a6b821bd0c58b94e registry.k8s.io/sig-storage/livenessprobe:v2.7.0],SizeBytes:8688564,},ContainerImage{Names:[registry.k8s.io/kas-network-proxy/proxy-agent@sha256:48f2a4ec3e10553a81b8dd1c6fa5fe4bcc9617f78e71c1ca89c6921335e2d7da registry.k8s.io/kas-network-proxy/proxy-agent:v0.0.33],SizeBytes:8512162,},ContainerImage{Names:[registry.k8s.io/networking/ingress-gce-404-server-with-metrics-amd64@sha256:7eb7b3cee4d33c10c49893ad3c386232b86d4067de5251294d4c620d6e072b93 registry.k8s.io/networking/ingress-gce-404-server-with-metrics-amd64:v1.10.11],SizeBytes:6463068,},ContainerImage{Names:[registry.k8s.io/metadata-proxy@sha256:e914645f22e946bce5165737e1b244e0a296ad1f0f81a9531adc57af2780978a registry.k8s.io/metadata-proxy:v0.1.12],SizeBytes:5301657,},ContainerImage{Names:[registry.k8s.io/e2e-test-images/busybox@sha256:c318242786b139d18676b1c09a0ad7f15fc17f8f16a5b2e625cd0dc8c9703daf registry.k8s.io/e2e-test-images/busybox:1.29-2],SizeBytes:732424,},ContainerImage{Names:[registry.k8s.io/e2e-test-images/busybox@sha256:2e0f836850e09b8b7cc937681d6194537a09fbd5f6b9e08f4d646a85128e8937 registry.k8s.io/e2e-test-images/busybox:1.29-4],SizeBytes:731990,},ContainerImage{Names:[registry.k8s.io/pause@sha256:7031c1b283388d2c2e09b57badb803c05ebed362dc88d84b480cc47f72a21097 registry.k8s.io/pause:3.9],SizeBytes:321520,},ContainerImage{Names:[registry.k8s.io/pause@sha256:9001185023633d17a2f98ff69b6ff2615b8ea02a825adffa40422f51dfdcde9d registry.k8s.io/pause:3.8],SizeBytes:311286,},},VolumesInUse:[kubernetes.io/csi/csi-hostpath-provisioning-6984^3b0457df-6d48-11ed-9e4d-d626cdd9a50a],VolumesAttached:[]AttachedVolume{AttachedVolume{Name:kubernetes.io/csi/csi-hostpath-provisioning-6984^3b0457df-6d48-11ed-9e4d-d626cdd9a50a,DevicePath:,},},Config:nil,},} Nov 26 05:15:40.103: INFO: Logging kubelet events for node bootstrap-e2e-minion-group-rwsn Nov 26 05:15:40.146: INFO: Logging pods the kubelet thinks is on node bootstrap-e2e-minion-group-rwsn Nov 26 05:15:40.214: INFO: kube-dns-autoscaler-5f6455f985-xktw8 started at 2022-11-26 05:05:25 +0000 UTC (0+1 container statuses recorded) Nov 26 05:15:40.214: INFO: Container autoscaler ready: false, restart count 5 Nov 26 05:15:40.214: INFO: coredns-6d97d5ddb-p2xpt started at 2022-11-26 05:05:31 +0000 UTC (0+1 container statuses recorded) Nov 26 05:15:40.214: INFO: Container coredns ready: false, restart count 6 Nov 26 05:15:40.214: INFO: csi-mockplugin-0 started at 2022-11-26 05:13:01 +0000 UTC (0+4 container statuses recorded) Nov 26 05:15:40.214: INFO: Container busybox ready: true, restart count 2 Nov 26 05:15:40.214: INFO: Container csi-provisioner ready: true, restart count 2 Nov 26 05:15:40.214: INFO: Container driver-registrar ready: true, restart count 2 Nov 26 05:15:40.214: INFO: Container mock ready: true, restart count 2 Nov 26 05:15:40.214: INFO: netserver-1 started at 2022-11-26 05:11:06 +0000 UTC (0+1 container statuses recorded) Nov 26 05:15:40.214: INFO: Container webserver ready: false, restart count 5 Nov 26 05:15:40.214: INFO: nfs-server started at 2022-11-26 05:13:35 +0000 UTC (0+1 container statuses recorded) Nov 26 05:15:40.214: INFO: Container nfs-server ready: true, restart count 0 Nov 26 05:15:40.214: INFO: kube-proxy-bootstrap-e2e-minion-group-rwsn started at 2022-11-26 05:05:15 +0000 UTC (0+1 container statuses recorded) Nov 26 05:15:40.214: INFO: Container kube-proxy ready: true, restart count 4 Nov 26 05:15:40.214: INFO: l7-default-backend-8549d69d99-8cfzv started at 2022-11-26 05:05:25 +0000 UTC (0+1 container statuses recorded) Nov 26 05:15:40.214: INFO: Container default-http-backend ready: true, restart count 0 Nov 26 05:15:40.214: INFO: volume-snapshot-controller-0 started at 2022-11-26 05:05:25 +0000 UTC (0+1 container statuses recorded) Nov 26 05:15:40.214: INFO: Container volume-snapshot-controller ready: false, restart count 4 Nov 26 05:15:40.214: INFO: csi-hostpathplugin-0 started at 2022-11-26 05:07:19 +0000 UTC (0+7 container statuses recorded) Nov 26 05:15:40.214: INFO: Container csi-attacher ready: false, restart count 6 Nov 26 05:15:40.214: INFO: Container csi-provisioner ready: false, restart count 6 Nov 26 05:15:40.214: INFO: Container csi-resizer ready: false, restart count 6 Nov 26 05:15:40.214: INFO: Container csi-snapshotter ready: false, restart count 6 Nov 26 05:15:40.214: INFO: Container hostpath ready: false, restart count 6 Nov 26 05:15:40.214: INFO: Container liveness-probe ready: false, restart count 6 Nov 26 05:15:40.214: INFO: Container node-driver-registrar ready: false, restart count 6 Nov 26 05:15:40.214: INFO: affinity-lb-esipp-transition-6h4g4 started at 2022-11-26 05:10:44 +0000 UTC (0+1 container statuses recorded) Nov 26 05:15:40.214: INFO: Container affinity-lb-esipp-transition ready: true, restart count 4 Nov 26 05:15:40.214: INFO: csi-mockplugin-0 started at 2022-11-26 05:08:11 +0000 UTC (0+4 container statuses recorded) Nov 26 05:15:40.214: INFO: Container busybox ready: true, restart count 3 Nov 26 05:15:40.214: INFO: Container csi-provisioner ready: true, restart count 3 Nov 26 05:15:40.214: INFO: Container driver-registrar ready: true, restart count 4 Nov 26 05:15:40.214: INFO: Container mock ready: true, restart count 4 Nov 26 05:15:40.214: INFO: netserver-1 started at 2022-11-26 05:11:29 +0000 UTC (0+1 container statuses recorded) Nov 26 05:15:40.214: INFO: Container webserver ready: true, restart count 5 Nov 26 05:15:40.214: INFO: pvc-volume-tester-5h2gx started at 2022-11-26 05:13:54 +0000 UTC (0+1 container statuses recorded) Nov 26 05:15:40.214: INFO: Container volume-tester ready: false, restart count 0 Nov 26 05:15:40.214: INFO: metadata-proxy-v0.1-qlckk started at 2022-11-26 05:05:16 +0000 UTC (0+2 container statuses recorded) Nov 26 05:15:40.214: INFO: Container metadata-proxy ready: true, restart count 0 Nov 26 05:15:40.214: INFO: Container prometheus-to-sd-exporter ready: true, restart count 0 Nov 26 05:15:40.214: INFO: netserver-1 started at 2022-11-26 05:09:51 +0000 UTC (0+1 container statuses recorded) Nov 26 05:15:40.214: INFO: Container webserver ready: true, restart count 5 Nov 26 05:15:40.214: INFO: csi-hostpathplugin-0 started at 2022-11-26 05:10:11 +0000 UTC (0+7 container statuses recorded) Nov 26 05:15:40.214: INFO: Container csi-attacher ready: false, restart count 5 Nov 26 05:15:40.214: INFO: Container csi-provisioner ready: false, restart count 5 Nov 26 05:15:40.214: INFO: Container csi-resizer ready: false, restart count 5 Nov 26 05:15:40.214: INFO: Container csi-snapshotter ready: false, restart count 5 Nov 26 05:15:40.214: INFO: Container hostpath ready: false, restart count 5 Nov 26 05:15:40.214: INFO: Container liveness-probe ready: false, restart count 5 Nov 26 05:15:40.214: INFO: Container node-driver-registrar ready: false, restart count 5 Nov 26 05:15:40.214: INFO: coredns-6d97d5ddb-xk6v4 started at 2022-11-26 05:05:25 +0000 UTC (0+1 container statuses recorded) Nov 26 05:15:40.214: INFO: Container coredns ready: false, restart count 5 Nov 26 05:15:40.214: INFO: konnectivity-agent-sxg5q started at 2022-11-26 05:05:25 +0000 UTC (0+1 container statuses recorded) Nov 26 05:15:40.214: INFO: Container konnectivity-agent ready: false, restart count 5 Nov 26 05:15:40.214: INFO: pod-subpath-test-dynamicpv-h6m7 started at 2022-11-26 05:07:31 +0000 UTC (1+1 container statuses recorded) Nov 26 05:15:40.214: INFO: Init container init-volume-dynamicpv-h6m7 ready: true, restart count 0 Nov 26 05:15:40.214: INFO: Container test-container-subpath-dynamicpv-h6m7 ready: false, restart count 0 Nov 26 05:15:40.755: INFO: Latency metrics for node bootstrap-e2e-minion-group-rwsn Nov 26 05:15:40.755: INFO: Logging node info for node bootstrap-e2e-minion-group-vhdj Nov 26 05:15:40.797: INFO: Node Info: &Node{ObjectMeta:{bootstrap-e2e-minion-group-vhdj 0d27fd16-9ec8-4428-bb1e-97fead6c7a66 6514 0 2022-11-26 05:05:23 +0000 UTC <nil> <nil> map[beta.kubernetes.io/arch:amd64 beta.kubernetes.io/instance-type:n1-standard-2 beta.kubernetes.io/os:linux cloud.google.com/metadata-proxy-ready:true failure-domain.beta.kubernetes.io/region:us-west1 failure-domain.beta.kubernetes.io/zone:us-west1-b kubernetes.io/arch:amd64 kubernetes.io/hostname:bootstrap-e2e-minion-group-vhdj kubernetes.io/os:linux node.kubernetes.io/instance-type:n1-standard-2 topology.hostpath.csi/node:bootstrap-e2e-minion-group-vhdj topology.kubernetes.io/region:us-west1 topology.kubernetes.io/zone:us-west1-b] map[csi.volume.kubernetes.io/nodeid:{"csi-hostpath-provisioning-7617":"bootstrap-e2e-minion-group-vhdj","csi-hostpath-volumeio-4786":"bootstrap-e2e-minion-group-vhdj"} node.alpha.kubernetes.io/ttl:0 volumes.kubernetes.io/controller-managed-attach-detach:true] [] [] [{kubelet Update v1 2022-11-26 05:05:23 +0000 UTC FieldsV1 {"f:metadata":{"f:annotations":{".":{},"f:volumes.kubernetes.io/controller-managed-attach-detach":{}},"f:labels":{".":{},"f:beta.kubernetes.io/arch":{},"f:beta.kubernetes.io/instance-type":{},"f:beta.kubernetes.io/os":{},"f:cloud.google.com/metadata-proxy-ready":{},"f:failure-domain.beta.kubernetes.io/region":{},"f:failure-domain.beta.kubernetes.io/zone":{},"f:kubernetes.io/arch":{},"f:kubernetes.io/hostname":{},"f:kubernetes.io/os":{},"f:node.kubernetes.io/instance-type":{},"f:topology.kubernetes.io/region":{},"f:topology.kubernetes.io/zone":{}}},"f:spec":{"f:providerID":{}}} } {kube-controller-manager Update v1 2022-11-26 05:05:24 +0000 UTC FieldsV1 {"f:metadata":{"f:annotations":{"f:node.alpha.kubernetes.io/ttl":{}}},"f:spec":{"f:podCIDR":{},"f:podCIDRs":{".":{},"v:\"10.64.3.0/24\"":{}}}} } {kube-controller-manager Update v1 2022-11-26 05:13:10 +0000 UTC FieldsV1 {"f:status":{"f:conditions":{"k:{\"type\":\"NetworkUnavailable\"}":{"f:lastHeartbeatTime":{},"f:lastTransitionTime":{},"f:message":{},"f:reason":{},"f:status":{}}}}} status} {kubelet Update v1 2022-11-26 05:15:24 +0000 UTC FieldsV1 {"f:metadata":{"f:annotations":{"f:csi.volume.kubernetes.io/nodeid":{}},"f:labels":{"f:topology.hostpath.csi/node":{}}},"f:status":{"f:conditions":{"k:{\"type\":\"DiskPressure\"}":{"f:lastHeartbeatTime":{}},"k:{\"type\":\"MemoryPressure\"}":{"f:lastHeartbeatTime":{}},"k:{\"type\":\"PIDPressure\"}":{"f:lastHeartbeatTime":{}},"k:{\"type\":\"Ready\"}":{"f:lastHeartbeatTime":{},"f:lastTransitionTime":{},"f:message":{},"f:reason":{},"f:status":{}}},"f:images":{}}} status} {node-problem-detector Update v1 2022-11-26 05:15:28 +0000 UTC FieldsV1 {"f:status":{"f:conditions":{"k:{\"type\":\"CorruptDockerOverlay2\"}":{".":{},"f:lastHeartbeatTime":{},"f:lastTransitionTime":{},"f:message":{},"f:reason":{},"f:status":{},"f:type":{}},"k:{\"type\":\"FrequentContainerdRestart\"}":{".":{},"f:lastHeartbeatTime":{},"f:lastTransitionTime":{},"f:message":{},"f:reason":{},"f:status":{},"f:type":{}},"k:{\"type\":\"FrequentDockerRestart\"}":{".":{},"f:lastHeartbeatTime":{},"f:lastTransitionTime":{},"f:message":{},"f:reason":{},"f:status":{},"f:type":{}},"k:{\"type\":\"FrequentKubeletRestart\"}":{".":{},"f:lastHeartbeatTime":{},"f:lastTransitionTime":{},"f:message":{},"f:reason":{},"f:status":{},"f:type":{}},"k:{\"type\":\"FrequentUnregisterNetDevice\"}":{".":{},"f:lastHeartbeatTime":{},"f:lastTransitionTime":{},"f:message":{},"f:reason":{},"f:status":{},"f:type":{}},"k:{\"type\":\"KernelDeadlock\"}":{".":{},"f:lastHeartbeatTime":{},"f:lastTransitionTime":{},"f:message":{},"f:reason":{},"f:status":{},"f:type":{}},"k:{\"type\":\"ReadonlyFilesystem\"}":{".":{},"f:lastHeartbeatTime":{},"f:lastTransitionTime":{},"f:message":{},"f:reason":{},"f:status":{},"f:type":{}}}}} status}]},Spec:NodeSpec{PodCIDR:10.64.3.0/24,DoNotUseExternalID:,ProviderID:gce://k8s-jkns-e2e-gce-ci-reboot/us-west1-b/bootstrap-e2e-minion-group-vhdj,Unschedulable:false,Taints:[]Taint{},ConfigSource:nil,PodCIDRs:[10.64.3.0/24],},Status:NodeStatus{Capacity:ResourceList{attachable-volumes-gce-pd: {{127 0} {<nil>} 127 DecimalSI},cpu: {{2 0} {<nil>} 2 DecimalSI},ephemeral-storage: {{101203873792 0} {<nil>} 98831908Ki BinarySI},hugepages-1Gi: {{0 0} {<nil>} 0 DecimalSI},hugepages-2Mi: {{0 0} {<nil>} 0 DecimalSI},memory: {{7815438336 0} {<nil>} BinarySI},pods: {{110 0} {<nil>} 110 DecimalSI},},Allocatable:ResourceList{attachable-volumes-gce-pd: {{127 0} {<nil>} 127 DecimalSI},cpu: {{2 0} {<nil>} 2 DecimalSI},ephemeral-storage: {{91083486262 0} {<nil>} 91083486262 DecimalSI},hugepages-1Gi: {{0 0} {<nil>} 0 DecimalSI},hugepages-2Mi: {{0 0} {<nil>} 0 DecimalSI},memory: {{7553294336 0} {<nil>} BinarySI},pods: {{110 0} {<nil>} 110 DecimalSI},},Phase:,Conditions:[]NodeCondition{NodeCondition{Type:ReadonlyFilesystem,Status:False,LastHeartbeatTime:2022-11-26 05:15:28 +0000 UTC,LastTransitionTime:2022-11-26 05:05:26 +0000 UTC,Reason:FilesystemIsNotReadOnly,Message:Filesystem is not read-only,},NodeCondition{Type:CorruptDockerOverlay2,Status:False,LastHeartbeatTime:2022-11-26 05:15:28 +0000 UTC,LastTransitionTime:2022-11-26 05:05:26 +0000 UTC,Reason:NoCorruptDockerOverlay2,Message:docker overlay2 is functioning properly,},NodeCondition{Type:FrequentUnregisterNetDevice,Status:False,LastHeartbeatTime:2022-11-26 05:15:28 +0000 UTC,LastTransitionTime:2022-11-26 05:05:26 +0000 UTC,Reason:NoFrequentUnregisterNetDevice,Message:node is functioning properly,},NodeCondition{Type:FrequentKubeletRestart,Status:False,LastHeartbeatTime:2022-11-26 05:15:28 +0000 UTC,LastTransitionTime:2022-11-26 05:05:26 +0000 UTC,Reason:NoFrequentKubeletRestart,Message:kubelet is functioning properly,},NodeCondition{Type:FrequentDockerRestart,Status:False,LastHeartbeatTime:2022-11-26 05:15:28 +0000 UTC,LastTransitionTime:2022-11-26 05:05:26 +0000 UTC,Reason:NoFrequentDockerRestart,Message:docker is functioning properly,},NodeCondition{Type:FrequentContainerdRestart,Status:False,LastHeartbeatTime:2022-11-26 05:15:28 +0000 UTC,LastTransitionTime:2022-11-26 05:05:26 +0000 UTC,Reason:NoFrequentContainerdRestart,Message:containerd is functioning properly,},NodeCondition{Type:KernelDeadlock,Status:False,LastHeartbeatTime:2022-11-26 05:15:28 +0000 UTC,LastTransitionTime:2022-11-26 05:05:26 +0000 UTC,Reason:KernelHasNoDeadlock,Message:kernel has no deadlock,},NodeCondition{Type:NetworkUnavailable,Status:False,LastHeartbeatTime:2022-11-26 05:05:39 +0000 UTC,LastTransitionTime:2022-11-26 05:05:39 +0000 UTC,Reason:RouteCreated,Message:RouteController created a route,},NodeCondition{Type:MemoryPressure,Status:False,LastHeartbeatTime:2022-11-26 05:13:20 +0000 UTC,LastTransitionTime:2022-11-26 05:05:23 +0000 UTC,Reason:KubeletHasSufficientMemory,Message:kubelet has sufficient memory available,},NodeCondition{Type:DiskPressure,Status:False,LastHeartbeatTime:2022-11-26 05:13:20 +0000 UTC,LastTransitionTime:2022-11-26 05:05:23 +0000 UTC,Reason:KubeletHasNoDiskPressure,Message:kubelet has no disk pressure,},NodeCondition{Type:PIDPressure,Status:False,LastHeartbeatTime:2022-11-26 05:13:20 +0000 UTC,LastTransitionTime:2022-11-26 05:05:23 +0000 UTC,Reason:KubeletHasSufficientPID,Message:kubelet has sufficient PID available,},NodeCondition{Type:Ready,Status:True,LastHeartbeatTime:2022-11-26 05:13:20 +0000 UTC,LastTransitionTime:2022-11-26 05:05:24 +0000 UTC,Reason:KubeletReady,Message:kubelet is posting ready status. AppArmor enabled,},},Addresses:[]NodeAddress{NodeAddress{Type:InternalIP,Address:10.138.0.3,},NodeAddress{Type:ExternalIP,Address:34.83.94.215,},NodeAddress{Type:InternalDNS,Address:bootstrap-e2e-minion-group-vhdj.c.k8s-jkns-e2e-gce-ci-reboot.internal,},NodeAddress{Type:Hostname,Address:bootstrap-e2e-minion-group-vhdj.c.k8s-jkns-e2e-gce-ci-reboot.internal,},},DaemonEndpoints:NodeDaemonEndpoints{KubeletEndpoint:DaemonEndpoint{Port:10250,},},NodeInfo:NodeSystemInfo{MachineID:31b0953ad3e740aeb7fc1a89edb5435a,SystemUUID:31b0953a-d3e7-40ae-b7fc-1a89edb5435a,BootID:14c054d0-bda9-4ddf-b9f6-e1742be33d25,KernelVersion:5.10.123+,OSImage:Container-Optimized OS from Google,ContainerRuntimeVersion:containerd://1.7.0-beta.0-149-gd06318622,KubeletVersion:v1.27.0-alpha.0.50+70617042976dc1,KubeProxyVersion:v1.27.0-alpha.0.50+70617042976dc1,OperatingSystem:linux,Architecture:amd64,},Images:[]ContainerImage{ContainerImage{Names:[registry.k8s.io/e2e-test-images/jessie-dnsutils@sha256:24aaf2626d6b27864c29de2097e8bbb840b3a414271bf7c8995e431e47d8408e registry.k8s.io/e2e-test-images/jessie-dnsutils:1.7],SizeBytes:112030336,},ContainerImage{Names:[registry.k8s.io/sig-storage/nfs-provisioner@sha256:e943bb77c7df05ebdc8c7888b2db289b13bf9f012d6a3a5a74f14d4d5743d439 registry.k8s.io/sig-storage/nfs-provisioner:v3.0.1],SizeBytes:90632047,},ContainerImage{Names:[registry.k8s.io/kube-proxy-amd64:v1.27.0-alpha.0.50_70617042976dc1],SizeBytes:67201736,},ContainerImage{Names:[registry.k8s.io/e2e-test-images/agnhost@sha256:16bbf38c463a4223d8cfe4da12bc61010b082a79b4bb003e2d3ba3ece5dd5f9e registry.k8s.io/e2e-test-images/agnhost:2.43],SizeBytes:51706353,},ContainerImage{Names:[gke.gcr.io/prometheus-to-sd@sha256:e739643c3939ba0b161425f45a1989eedfc4a3b166db9a7100863296b4c70510 gke.gcr.io/prometheus-to-sd:v0.11.1-gke.1],SizeBytes:48742566,},ContainerImage{Names:[registry.k8s.io/sig-storage/csi-provisioner@sha256:ee3b525d5b89db99da3b8eb521d9cd90cb6e9ef0fbb651e98bb37be78d36b5b8 registry.k8s.io/sig-storage/csi-provisioner:v3.3.0],SizeBytes:25491225,},ContainerImage{Names:[registry.k8s.io/sig-storage/csi-resizer@sha256:425d8f1b769398127767b06ed97ce62578a3179bcb99809ce93a1649e025ffe7 registry.k8s.io/sig-storage/csi-resizer:v1.6.0],SizeBytes:24148884,},ContainerImage{Names:[registry.k8s.io/sig-storage/csi-snapshotter@sha256:291334908ddf71a4661fd7f6d9d97274de8a5378a2b6fdfeb2ce73414a34f82f registry.k8s.io/sig-storage/csi-snapshotter:v6.1.0],SizeBytes:23881995,},ContainerImage{Names:[registry.k8s.io/sig-storage/csi-attacher@sha256:9a685020911e2725ad019dbce6e4a5ab93d51e3d4557f115e64343345e05781b registry.k8s.io/sig-storage/csi-attacher:v4.0.0],SizeBytes:23847201,},ContainerImage{Names:[registry.k8s.io/sig-storage/hostpathplugin@sha256:92257881c1d6493cf18299a24af42330f891166560047902b8d431fb66b01af5 registry.k8s.io/sig-storage/hostpathplugin:v1.9.0],SizeBytes:18758628,},ContainerImage{Names:[registry.k8s.io/sig-storage/csi-node-driver-registrar@sha256:0103eee7c35e3e0b5cd8cdca9850dc71c793cdeb6669d8be7a89440da2d06ae4 registry.k8s.io/sig-storage/csi-node-driver-registrar:v2.5.1],SizeBytes:9133109,},ContainerImage{Names:[registry.k8s.io/sig-storage/livenessprobe@sha256:933940f13b3ea0abc62e656c1aa5c5b47c04b15d71250413a6b821bd0c58b94e registry.k8s.io/sig-storage/livenessprobe:v2.7.0],SizeBytes:8688564,},ContainerImage{Names:[registry.k8s.io/kas-network-proxy/proxy-agent@sha256:48f2a4ec3e10553a81b8dd1c6fa5fe4bcc9617f78e71c1ca89c6921335e2d7da registry.k8s.io/kas-network-proxy/proxy-agent:v0.0.33],SizeBytes:8512162,},ContainerImage{Names:[registry.k8s.io/metadata-proxy@sha256:e914645f22e946bce5165737e1b244e0a296ad1f0f81a9531adc57af2780978a registry.k8s.io/metadata-proxy:v0.1.12],SizeBytes:5301657,},ContainerImage{Names:[registry.k8s.io/e2e-test-images/busybox@sha256:2e0f836850e09b8b7cc937681d6194537a09fbd5f6b9e08f4d646a85128e8937 registry.k8s.io/e2e-test-images/busybox:1.29-4],SizeBytes:731990,},ContainerImage{Names:[registry.k8s.io/pause@sha256:9001185023633d17a2f98ff69b6ff2615b8ea02a825adffa40422f51dfdcde9d registry.k8s.io/pause:3.8],SizeBytes:311286,},},VolumesInUse:[],VolumesAttached:[]AttachedVolume{},Config:nil,},} Nov 26 05:15:40.797: INFO: Logging kubelet events for node bootstrap-e2e-minion-group-vhdj Nov 26 05:15:40.846: INFO: Logging pods the kubelet thinks is on node bootstrap-e2e-minion-group-vhdj Nov 26 05:15:40.913: INFO: affinity-lb-esipp-transition-ml7rt started at 2022-11-26 05:10:44 +0000 UTC (0+1 container statuses recorded) Nov 26 05:15:40.913: INFO: Container affinity-lb-esipp-transition ready: true, restart count 2 Nov 26 05:15:40.913: INFO: netserver-2 started at 2022-11-26 05:09:51 +0000 UTC (0+1 container statuses recorded) Nov 26 05:15:40.913: INFO: Container webserver ready: true, restart count 3 Nov 26 05:15:40.913: INFO: csi-hostpathplugin-0 started at 2022-11-26 05:10:50 +0000 UTC (0+7 container statuses recorded) Nov 26 05:15:40.913: INFO: Container csi-attacher ready: true, restart count 1 Nov 26 05:15:40.913: INFO: Container csi-provisioner ready: true, restart count 1 Nov 26 05:15:40.913: INFO: Container csi-resizer ready: true, restart count 1 Nov 26 05:15:40.913: INFO: Container csi-snapshotter ready: true, restart count 1 Nov 26 05:15:40.913: INFO: Container hostpath ready: true, restart count 1 Nov 26 05:15:40.913: INFO: Container liveness-probe ready: true, restart count 1 Nov 26 05:15:40.913: INFO: Container node-driver-registrar ready: true, restart count 1 Nov 26 05:15:40.913: INFO: test-container-pod started at 2022-11-26 05:12:10 +0000 UTC (0+1 container statuses recorded) Nov 26 05:15:40.913: INFO: Container webserver ready: true, restart count 0 Nov 26 05:15:40.913: INFO: back-off-cap started at 2022-11-26 05:07:40 +0000 UTC (0+1 container statuses recorded) Nov 26 05:15:40.913: INFO: Container back-off-cap ready: false, restart count 6 Nov 26 05:15:40.913: INFO: csi-hostpathplugin-0 started at 2022-11-26 05:13:07 +0000 UTC (0+7 container statuses recorded) Nov 26 05:15:40.913: INFO: Container csi-attacher ready: true, restart count 1 Nov 26 05:15:40.913: INFO: Container csi-provisioner ready: true, restart count 1 Nov 26 05:15:40.913: INFO: Container csi-resizer ready: true, restart count 1 Nov 26 05:15:40.913: INFO: Container csi-snapshotter ready: true, restart count 1 Nov 26 05:15:40.913: INFO: Container hostpath ready: true, restart count 1 Nov 26 05:15:40.913: INFO: Container liveness-probe ready: true, restart count 1 Nov 26 05:15:40.913: INFO: Container node-driver-registrar ready: true, restart count 1 Nov 26 05:15:40.913: INFO: konnectivity-agent-nt4fd started at 2022-11-26 05:05:39 +0000 UTC (0+1 container statuses recorded) Nov 26 05:15:40.913: INFO: Container konnectivity-agent ready: true, restart count 5 Nov 26 05:15:40.913: INFO: netserver-2 started at 2022-11-26 05:11:06 +0000 UTC (0+1 container statuses recorded) Nov 26 05:15:40.913: INFO: Container webserver ready: true, restart count 1 Nov 26 05:15:40.913: INFO: netserver-2 started at 2022-11-26 05:11:29 +0000 UTC (0+1 container statuses recorded) Nov 26 05:15:40.913: INFO: Container webserver ready: true, restart count 2 Nov 26 05:15:40.913: INFO: pod-9fe7e984-9a6a-4bb0-9c65-0628a6a2f792 started at 2022-11-26 05:08:13 +0000 UTC (0+1 container statuses recorded) Nov 26 05:15:40.913: INFO: Container write-pod ready: false, restart count 0 Nov 26 05:15:40.913: INFO: kube-proxy-bootstrap-e2e-minion-group-vhdj started at 2022-11-26 05:05:23 +0000 UTC (0+1 container statuses recorded) Nov 26 05:15:40.913: INFO: Container kube-proxy ready: false, restart count 5 Nov 26 05:15:40.913: INFO: pod-subpath-test-dynamicpv-4nfp started at 2022-11-26 05:11:17 +0000 UTC (1+2 container statuses recorded) Nov 26 05:15:40.913: INFO: Init container init-volume-dynamicpv-4nfp ready: true, restart count 0 Nov 26 05:15:40.913: INFO: Container test-container-subpath-dynamicpv-4nfp ready: false, restart count 2 Nov 26 05:15:40.913: INFO: Container test-container-volume-dynamicpv-4nfp ready: false, restart count 2 Nov 26 05:15:40.913: INFO: forbid-27823992-j66hf started at 2022-11-26 05:12:10 +0000 UTC (0+1 container statuses recorded) Nov 26 05:15:40.913: INFO: Container c ready: true, restart count 1 Nov 26 05:15:40.913: INFO: mutability-test-55dcq started at 2022-11-26 05:13:49 +0000 UTC (0+1 container statuses recorded) Nov 26 05:15:40.913: INFO: Container netexec ready: true, restart count 1 Nov 26 05:15:40.913: INFO: metadata-proxy-v0.1-wn5l5 started at 2022-11-26 05:05:24 +0000 UTC (0+2 container statuses recorded) Nov 26 05:15:40.913: INFO: Container metadata-proxy ready: true, restart count 0 Nov 26 05:15:40.913: INFO: Container prometheus-to-sd-exporter ready: true, restart count 0 Nov 26 05:15:41.144: INFO: Latency metrics for node bootstrap-e2e-minion-group-vhdj [DeferCleanup (Each)] [sig-auth] ServiceAccounts tear down framework | framework.go:193 STEP: Destroying namespace "svcaccounts-6307" for this suite. 11/26/22 05:15:41.144
Filter through log files | View test history on testgrid
go run hack/e2e.go -v --test --test_args='--ginkgo.focus=Kubernetes\se2e\ssuite\s\[It\]\s\[sig\-cloud\-provider\-gcp\]\sAddon\supdate\sshould\spropagate\sadd\-on\sfile\schanges\s\[Slow\]$'
test/e2e/cloud/gcp/addon_update.go:349 k8s.io/kubernetes/test/e2e/cloud/gcp.waitForServiceInAddonTest({0x801de88?, 0xc00220d860?}, {0x75ce977?, 0x2?}, {0x7631374?, 0x7fec6c0?}, 0xb8?) test/e2e/cloud/gcp/addon_update.go:349 +0x54 k8s.io/kubernetes/test/e2e/cloud/gcp.glob..func1.3() test/e2e/cloud/gcp/addon_update.go:336 +0x18c7 There were additional failures detected after the initial failure: [FAILED] Nov 26 05:31:45.686: failed to list events in namespace "addon-update-test-2109": Get "https://35.227.179.144/api/v1/namespaces/addon-update-test-2109/events": dial tcp 35.227.179.144:443: connect: connection refused In [DeferCleanup (Each)] at: test/e2e/framework/debug/dump.go:44 ---------- [FAILED] Nov 26 05:31:45.726: Couldn't delete ns: "addon-update-test-2109": Delete "https://35.227.179.144/api/v1/namespaces/addon-update-test-2109": dial tcp 35.227.179.144:443: connect: connection refused (&url.Error{Op:"Delete", URL:"https://35.227.179.144/api/v1/namespaces/addon-update-test-2109", Err:(*net.OpError)(0xc00222e280)}) In [DeferCleanup (Each)] at: test/e2e/framework/framework.go:370from junit_01.xml
[BeforeEach] [sig-cloud-provider-gcp] Addon update set up framework | framework.go:178 STEP: Creating a kubernetes client 11/26/22 05:27:43.118 Nov 26 05:27:43.118: INFO: >>> kubeConfig: /workspace/.kube/config STEP: Building a namespace api object, basename addon-update-test 11/26/22 05:27:43.12 STEP: Waiting for a default service account to be provisioned in namespace 11/26/22 05:29:29.042 STEP: Waiting for kube-root-ca.crt to be provisioned in namespace 11/26/22 05:29:29.128 [BeforeEach] [sig-cloud-provider-gcp] Addon update test/e2e/framework/metrics/init/init.go:31 [BeforeEach] [sig-cloud-provider-gcp] Addon update test/e2e/cloud/gcp/addon_update.go:223 [It] should propagate add-on file changes [Slow] test/e2e/cloud/gcp/addon_update.go:244 Nov 26 05:29:33.605: INFO: Executing 'mkdir -p addon-test-dir/addon-update-test-2109' on 35.227.179.144:22 Nov 26 05:29:33.809: INFO: Writing remote file 'addon-test-dir/addon-update-test-2109/addon-reconcile-controller.yaml' on 35.227.179.144:22 Nov 26 05:29:33.927: INFO: Writing remote file 'addon-test-dir/addon-update-test-2109/addon-reconcile-controller-Updated.yaml' on 35.227.179.144:22 Nov 26 05:29:34.044: INFO: Writing remote file 'addon-test-dir/addon-update-test-2109/addon-deprecated-label-service.yaml' on 35.227.179.144:22 Nov 26 05:29:34.162: INFO: Writing remote file 'addon-test-dir/addon-update-test-2109/addon-deprecated-label-service-updated.yaml' on 35.227.179.144:22 Nov 26 05:29:34.280: INFO: Writing remote file 'addon-test-dir/addon-update-test-2109/addon-ensure-exists-service.yaml' on 35.227.179.144:22 Nov 26 05:29:34.398: INFO: Writing remote file 'addon-test-dir/addon-update-test-2109/addon-ensure-exists-service-updated.yaml' on 35.227.179.144:22 Nov 26 05:29:34.516: INFO: Writing remote file 'addon-test-dir/addon-update-test-2109/invalid-addon-controller.yaml' on 35.227.179.144:22 Nov 26 05:29:34.633: INFO: Executing 'sudo rm -rf /etc/kubernetes/addons/addon-test-dir' on 35.227.179.144:22 Nov 26 05:29:34.736: INFO: Executing 'sudo mkdir -p /etc/kubernetes/addons/addon-test-dir/addon-update-test-2109' on 35.227.179.144:22 STEP: copy invalid manifests to the destination dir 11/26/22 05:29:34.827 Nov 26 05:29:34.827: INFO: Executing 'sudo cp addon-test-dir/addon-update-test-2109/invalid-addon-controller.yaml /etc/kubernetes/addons/addon-test-dir/addon-update-test-2109/invalid-addon-controller.yaml' on 35.227.179.144:22 STEP: copy new manifests 11/26/22 05:29:34.919 Nov 26 05:29:34.919: INFO: Executing 'sudo cp addon-test-dir/addon-update-test-2109/addon-reconcile-controller.yaml /etc/kubernetes/addons/addon-test-dir/addon-update-test-2109/addon-reconcile-controller.yaml' on 35.227.179.144:22 Nov 26 05:29:35.010: INFO: Executing 'sudo cp addon-test-dir/addon-update-test-2109/addon-deprecated-label-service.yaml /etc/kubernetes/addons/addon-test-dir/addon-update-test-2109/addon-deprecated-label-service.yaml' on 35.227.179.144:22 Nov 26 05:29:35.099: INFO: Executing 'sudo cp addon-test-dir/addon-update-test-2109/addon-ensure-exists-service.yaml /etc/kubernetes/addons/addon-test-dir/addon-update-test-2109/addon-ensure-exists-service.yaml' on 35.227.179.144:22 Nov 26 05:29:35.262: INFO: Get ReplicationController addon-reconcile-test in namespace kube-system failed (replicationcontrollers "addon-reconcile-test" not found). Nov 26 05:29:38.330: INFO: Get ReplicationController addon-reconcile-test in namespace kube-system failed (replicationcontrollers "addon-reconcile-test" not found). Nov 26 05:29:41.328: INFO: Get ReplicationController addon-reconcile-test in namespace kube-system failed (replicationcontrollers "addon-reconcile-test" not found). Nov 26 05:29:44.314: INFO: Get ReplicationController addon-reconcile-test in namespace kube-system failed (replicationcontrollers "addon-reconcile-test" not found). Nov 26 05:29:47.348: INFO: Get ReplicationController addon-reconcile-test in namespace kube-system failed (replicationcontrollers "addon-reconcile-test" not found). Nov 26 05:29:50.357: INFO: Get ReplicationController addon-reconcile-test in namespace kube-system failed (replicationcontrollers "addon-reconcile-test" not found). Nov 26 05:29:53.393: INFO: Get ReplicationController addon-reconcile-test in namespace kube-system failed (replicationcontrollers "addon-reconcile-test" not found). Nov 26 05:29:56.320: INFO: Get ReplicationController addon-reconcile-test in namespace kube-system failed (replicationcontrollers "addon-reconcile-test" not found). Nov 26 05:29:59.446: INFO: Get ReplicationController addon-reconcile-test in namespace kube-system failed (replicationcontrollers "addon-reconcile-test" not found). Nov 26 05:30:02.328: INFO: Get ReplicationController addon-reconcile-test in namespace kube-system failed (replicationcontrollers "addon-reconcile-test" not found). Nov 26 05:30:05.492: INFO: Get ReplicationController addon-reconcile-test in namespace kube-system failed (replicationcontrollers "addon-reconcile-test" not found). Nov 26 05:30:08.354: INFO: Get ReplicationController addon-reconcile-test in namespace kube-system failed (replicationcontrollers "addon-reconcile-test" not found). Nov 26 05:30:11.349: INFO: Get ReplicationController addon-reconcile-test in namespace kube-system failed (replicationcontrollers "addon-reconcile-test" not found). Nov 26 05:30:14.361: INFO: Get ReplicationController addon-reconcile-test in namespace kube-system failed (replicationcontrollers "addon-reconcile-test" not found). Nov 26 05:30:17.353: INFO: Get ReplicationController addon-reconcile-test in namespace kube-system failed (replicationcontrollers "addon-reconcile-test" not found). Nov 26 05:30:20.335: INFO: Get ReplicationController addon-reconcile-test in namespace kube-system failed (replicationcontrollers "addon-reconcile-test" not found). Nov 26 05:30:23.379: INFO: Get ReplicationController addon-reconcile-test in namespace kube-system failed (replicationcontrollers "addon-reconcile-test" not found). Nov 26 05:30:26.332: INFO: ReplicationController addon-reconcile-test in namespace kube-system found. Nov 26 05:30:26.378: INFO: Service addon-deprecated-label-test in namespace kube-system found. Nov 26 05:30:26.432: INFO: Service addon-ensure-exists-test in namespace kube-system found. STEP: update manifests 11/26/22 05:30:26.432 Nov 26 05:30:26.432: INFO: Executing 'sudo cp addon-test-dir/addon-update-test-2109/addon-reconcile-controller-Updated.yaml /etc/kubernetes/addons/addon-test-dir/addon-update-test-2109/addon-reconcile-controller.yaml' on 35.227.179.144:22 Nov 26 05:30:26.535: INFO: Executing 'sudo cp addon-test-dir/addon-update-test-2109/addon-deprecated-label-service-updated.yaml /etc/kubernetes/addons/addon-test-dir/addon-update-test-2109/addon-deprecated-label-service.yaml' on 35.227.179.144:22 Nov 26 05:30:26.627: INFO: Executing 'sudo cp addon-test-dir/addon-update-test-2109/addon-ensure-exists-service-updated.yaml /etc/kubernetes/addons/addon-test-dir/addon-update-test-2109/addon-ensure-exists-service.yaml' on 35.227.179.144:22 Nov 26 05:30:26.786: INFO: ReplicationController with newLabel=addon-reconcile-test in namespace kube-system disappeared. Nov 26 05:30:29.853: INFO: ReplicationController with newLabel=addon-reconcile-test in namespace kube-system disappeared. Nov 26 05:30:32.845: INFO: ReplicationController with newLabel=addon-reconcile-test in namespace kube-system disappeared. Nov 26 05:30:35.836: INFO: ReplicationController with newLabel=addon-reconcile-test in namespace kube-system disappeared. Nov 26 05:30:38.839: INFO: ReplicationController with newLabel=addon-reconcile-test in namespace kube-system disappeared. Nov 26 05:30:41.845: INFO: ReplicationController with newLabel=addon-reconcile-test in namespace kube-system disappeared. Nov 26 05:30:44.841: INFO: ReplicationController with newLabel=addon-reconcile-test in namespace kube-system disappeared. Nov 26 05:30:47.839: INFO: ReplicationController with newLabel=addon-reconcile-test in namespace kube-system disappeared. Nov 26 05:30:50.845: INFO: ReplicationController with newLabel=addon-reconcile-test in namespace kube-system disappeared. Nov 26 05:30:53.893: INFO: ReplicationController with newLabel=addon-reconcile-test in namespace kube-system disappeared. Nov 26 05:30:56.845: INFO: ReplicationController with newLabel=addon-reconcile-test in namespace kube-system disappeared. Nov 26 05:30:59.869: INFO: ReplicationController with newLabel=addon-reconcile-test in namespace kube-system disappeared. Nov 26 05:31:02.887: INFO: ReplicationController with newLabel=addon-reconcile-test in namespace kube-system disappeared. Nov 26 05:31:05.840: INFO: ReplicationController with newLabel=addon-reconcile-test in namespace kube-system disappeared. Nov 26 05:31:08.841: INFO: ReplicationController with newLabel=addon-reconcile-test in namespace kube-system disappeared. Nov 26 05:31:11.880: INFO: ReplicationController with newLabel=addon-reconcile-test in namespace kube-system disappeared. Nov 26 05:31:14.845: INFO: ReplicationController with newLabel=addon-reconcile-test in namespace kube-system disappeared. Nov 26 05:31:17.849: INFO: ReplicationController with newLabel=addon-reconcile-test in namespace kube-system disappeared. Nov 26 05:31:20.840: INFO: ReplicationController with newLabel=addon-reconcile-test in namespace kube-system disappeared. Nov 26 05:31:23.912: INFO: ReplicationController with newLabel=addon-reconcile-test in namespace kube-system disappeared. Nov 26 05:31:26.866: INFO: ReplicationController with newLabel=addon-reconcile-test in namespace kube-system disappeared. Nov 26 05:31:29.837: INFO: ReplicationController with newLabel=addon-reconcile-test in namespace kube-system found. Nov 26 05:31:29.906: INFO: Service with newLabel=addon-deprecated-label-test in namespace kube-system found. Nov 26 05:31:29.978: INFO: Service with newLabel=addon-ensure-exists-test in namespace kube-system disappeared. STEP: remove manifests 11/26/22 05:31:29.978 Nov 26 05:31:29.978: INFO: Executing 'sudo rm /etc/kubernetes/addons/addon-test-dir/addon-update-test-2109/addon-reconcile-controller.yaml' on 35.227.179.144:22 Nov 26 05:31:30.069: INFO: Executing 'sudo rm /etc/kubernetes/addons/addon-test-dir/addon-update-test-2109/addon-deprecated-label-service.yaml' on 35.227.179.144:22 Nov 26 05:31:30.160: INFO: Executing 'sudo rm /etc/kubernetes/addons/addon-test-dir/addon-update-test-2109/addon-ensure-exists-service.yaml' on 35.227.179.144:22 Nov 26 05:31:30.315: INFO: ReplicationController addon-reconcile-test in namespace kube-system found. Nov 26 05:31:33.484: INFO: ReplicationController addon-reconcile-test in namespace kube-system found. Nov 26 05:31:36.408: INFO: ReplicationController addon-reconcile-test in namespace kube-system found. Nov 26 05:31:39.380: INFO: ReplicationController addon-reconcile-test in namespace kube-system found. Nov 26 05:31:42.365: INFO: ReplicationController addon-reconcile-test in namespace kube-system found. Nov 26 05:31:45.355: INFO: Get ReplicationController addon-reconcile-test in namespace kube-system failed (Get "https://35.227.179.144/api/v1/namespaces/kube-system/replicationcontrollers/addon-reconcile-test": dial tcp 35.227.179.144:443: connect: connection refused). Nov 26 05:31:45.394: INFO: Non-retryable failure while getting service. Nov 26 05:31:45.394: INFO: Unexpected error: <*errors.errorString | 0xc0014ce470>: { s: "error waiting for service kube-system/addon-deprecated-label-test to disappear: Get \"https://35.227.179.144/api/v1/namespaces/kube-system/services/addon-deprecated-label-test\": dial tcp 35.227.179.144:443: connect: connection refused", } Nov 26 05:31:45.394: FAIL: error waiting for service kube-system/addon-deprecated-label-test to disappear: Get "https://35.227.179.144/api/v1/namespaces/kube-system/services/addon-deprecated-label-test": dial tcp 35.227.179.144:443: connect: connection refused Full Stack Trace k8s.io/kubernetes/test/e2e/cloud/gcp.waitForServiceInAddonTest({0x801de88?, 0xc00220d860?}, {0x75ce977?, 0x2?}, {0x7631374?, 0x7fec6c0?}, 0xb8?) test/e2e/cloud/gcp/addon_update.go:349 +0x54 k8s.io/kubernetes/test/e2e/cloud/gcp.glob..func1.3() test/e2e/cloud/gcp/addon_update.go:336 +0x18c7 Nov 26 05:31:45.394: INFO: Cleaning up ensure exist class addon. Nov 26 05:31:45.434: INFO: Unexpected error: <*url.Error | 0xc004be8420>: { Op: "Delete", URL: "https://35.227.179.144/api/v1/namespaces/kube-system/services/addon-ensure-exists-test", Err: <*net.OpError | 0xc003d6c5f0>{ Op: "dial", Net: "tcp", Source: nil, Addr: <*net.TCPAddr | 0xc00250eba0>{ IP: [0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 255, 255, 35, 227, 179, 144], Port: 443, Zone: "", }, Err: <*os.SyscallError | 0xc000d12220>{ Syscall: "connect", Err: <syscall.Errno>0x6f, }, }, } Nov 26 05:31:45.434: FAIL: Delete "https://35.227.179.144/api/v1/namespaces/kube-system/services/addon-ensure-exists-test": dial tcp 35.227.179.144:443: connect: connection refused Full Stack Trace k8s.io/kubernetes/test/e2e/cloud/gcp.glob..func1.3.1() test/e2e/cloud/gcp/addon_update.go:308 +0xe5 panic({0x70eb7e0, 0xc00011cf50}) /usr/local/go/src/runtime/panic.go:884 +0x212 k8s.io/kubernetes/test/e2e/framework.Fail({0xc0000925a0, 0xe9}, {0xc0006cf7a8?, 0xc0000925a0?, 0xc0006cf7d0?}) test/e2e/framework/log.go:61 +0x145 k8s.io/kubernetes/test/e2e/framework.ExpectNoErrorWithOffset(0x1, {0x7fa3ee0, 0xc0014ce470}, {0x0?, 0x7631374?, 0x1b?}) test/e2e/framework/expect.go:76 +0x267 k8s.io/kubernetes/test/e2e/framework.ExpectNoError(...) test/e2e/framework/expect.go:43 k8s.io/kubernetes/test/e2e/cloud/gcp.waitForServiceInAddonTest({0x801de88?, 0xc00220d860?}, {0x75ce977?, 0x2?}, {0x7631374?, 0x7fec6c0?}, 0xb8?) test/e2e/cloud/gcp/addon_update.go:349 +0x54 k8s.io/kubernetes/test/e2e/cloud/gcp.glob..func1.3() test/e2e/cloud/gcp/addon_update.go:336 +0x18c7 Nov 26 05:31:45.434: INFO: Executing 'sudo rm -rf /etc/kubernetes/addons/addon-test-dir' on 35.227.179.144:22 Nov 26 05:31:45.521: INFO: Executing 'rm -rf addon-test-dir' on 35.227.179.144:22 [AfterEach] [sig-cloud-provider-gcp] Addon update test/e2e/framework/node/init/init.go:32 Nov 26 05:31:45.607: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready [AfterEach] [sig-cloud-provider-gcp] Addon update test/e2e/cloud/gcp/addon_update.go:237 [DeferCleanup (Each)] [sig-cloud-provider-gcp] Addon update test/e2e/framework/metrics/init/init.go:33 [DeferCleanup (Each)] [sig-cloud-provider-gcp] Addon update dump namespaces | framework.go:196 STEP: dump namespace information after failure 11/26/22 05:31:45.646 STEP: Collecting events from namespace "addon-update-test-2109". 11/26/22 05:31:45.646 Nov 26 05:31:45.685: INFO: Unexpected error: failed to list events in namespace "addon-update-test-2109": <*url.Error | 0xc0044e8000>: { Op: "Get", URL: "https://35.227.179.144/api/v1/namespaces/addon-update-test-2109/events", Err: <*net.OpError | 0xc004a6e000>{ Op: "dial", Net: "tcp", Source: nil, Addr: <*net.TCPAddr | 0xc004be8d80>{ IP: [0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 255, 255, 35, 227, 179, 144], Port: 443, Zone: "", }, Err: <*os.SyscallError | 0xc0005202e0>{ Syscall: "connect", Err: <syscall.Errno>0x6f, }, }, } Nov 26 05:31:45.686: FAIL: failed to list events in namespace "addon-update-test-2109": Get "https://35.227.179.144/api/v1/namespaces/addon-update-test-2109/events": dial tcp 35.227.179.144:443: connect: connection refused Full Stack Trace k8s.io/kubernetes/test/e2e/framework/debug.dumpEventsInNamespace(0xc004d305c0, {0xc00449e378, 0x16}) test/e2e/framework/debug/dump.go:44 +0x191 k8s.io/kubernetes/test/e2e/framework/debug.DumpAllNamespaceInfo({0x801de88, 0xc00220d860}, {0xc00449e378, 0x16}) test/e2e/framework/debug/dump.go:62 +0x8d k8s.io/kubernetes/test/e2e/framework/debug/init.init.0.func1.1(0xc004d30650?, {0xc00449e378?, 0x7fa7740?}) test/e2e/framework/debug/init/init.go:34 +0x32 k8s.io/kubernetes/test/e2e/framework.(*Framework).dumpNamespaceInfo.func1() test/e2e/framework/framework.go:274 +0x6d k8s.io/kubernetes/test/e2e/framework.(*Framework).dumpNamespaceInfo(0xc0010d3d10) test/e2e/framework/framework.go:271 +0x179 reflect.Value.call({0x6627cc0?, 0xc0040ddb40?, 0xae73300?}, {0x75b6e72, 0x4}, {0xae73300, 0x0, 0x3cf0a1a?}) /usr/local/go/src/reflect/value.go:584 +0x8c5 reflect.Value.Call({0x6627cc0?, 0xc0040ddb40?, 0x1000000010000?}, {0xae73300?, 0xc001e0f690?, 0xc003b3bf70?}) /usr/local/go/src/reflect/value.go:368 +0xbc [DeferCleanup (Each)] [sig-cloud-provider-gcp] Addon update tear down framework | framework.go:193 STEP: Destroying namespace "addon-update-test-2109" for this suite. 11/26/22 05:31:45.686 Nov 26 05:31:45.725: FAIL: Couldn't delete ns: "addon-update-test-2109": Delete "https://35.227.179.144/api/v1/namespaces/addon-update-test-2109": dial tcp 35.227.179.144:443: connect: connection refused (&url.Error{Op:"Delete", URL:"https://35.227.179.144/api/v1/namespaces/addon-update-test-2109", Err:(*net.OpError)(0xc00222e280)}) Full Stack Trace k8s.io/kubernetes/test/e2e/framework.(*Framework).AfterEach.func1() test/e2e/framework/framework.go:370 +0x4fe k8s.io/kubernetes/test/e2e/framework.(*Framework).AfterEach(0xc0010d3d10) test/e2e/framework/framework.go:383 +0x1ca reflect.Value.call({0x6627cc0?, 0xc0040ddac0?, 0x0?}, {0x75b6e72, 0x4}, {0xae73300, 0x0, 0x0?}) /usr/local/go/src/reflect/value.go:584 +0x8c5 reflect.Value.Call({0x6627cc0?, 0xc0040ddac0?, 0x0?}, {0xae73300?, 0x0?, 0x0?}) /usr/local/go/src/reflect/value.go:368 +0xbc
Filter through log files | View test history on testgrid
go run hack/e2e.go -v --test --test_args='--ginkgo.focus=Kubernetes\se2e\ssuite\s\[It\]\s\[sig\-network\]\sLoadBalancers\sESIPP\s\[Slow\]\sshould\shandle\supdates\sto\sExternalTrafficPolicy\sfield$'
test/e2e/framework/network/utils.go:866 k8s.io/kubernetes/test/e2e/framework/network.(*NetworkingTestConfig).createNetProxyPods(0xc001006b60, {0x75c6f7c, 0x9}, 0xc0027b59e0) test/e2e/framework/network/utils.go:866 +0x1d0 k8s.io/kubernetes/test/e2e/framework/network.(*NetworkingTestConfig).setupCore(0xc001006b60, 0x7faf5c713740?) test/e2e/framework/network/utils.go:763 +0x55 k8s.io/kubernetes/test/e2e/framework/network.(*NetworkingTestConfig).setup(0xc001006b60, 0x3c?) test/e2e/framework/network/utils.go:778 +0x3e k8s.io/kubernetes/test/e2e/framework/network.NewNetworkingTestConfig(0xc001228000, {0x0, 0x0, 0x7f8f6d0?}) test/e2e/framework/network/utils.go:131 +0x125 k8s.io/kubernetes/test/e2e/network.glob..func20.7() test/e2e/network/loadbalancer.go:1544 +0x417 There were additional failures detected after the initial failure: [FAILED] Nov 26 05:13:58.009: failed to list events in namespace "esipp-2673": Get "https://35.227.179.144/api/v1/namespaces/esipp-2673/events": dial tcp 35.227.179.144:443: connect: connection refused In [DeferCleanup (Each)] at: test/e2e/framework/debug/dump.go:44 ---------- [FAILED] Nov 26 05:13:58.049: Couldn't delete ns: "esipp-2673": Delete "https://35.227.179.144/api/v1/namespaces/esipp-2673": dial tcp 35.227.179.144:443: connect: connection refused (&url.Error{Op:"Delete", URL:"https://35.227.179.144/api/v1/namespaces/esipp-2673", Err:(*net.OpError)(0xc0018e0d20)}) In [DeferCleanup (Each)] at: test/e2e/framework/framework.go:370from junit_01.xml
[BeforeEach] [sig-network] LoadBalancers ESIPP [Slow] set up framework | framework.go:178 STEP: Creating a kubernetes client 11/26/22 05:09:50.538 Nov 26 05:09:50.538: INFO: >>> kubeConfig: /workspace/.kube/config STEP: Building a namespace api object, basename esipp 11/26/22 05:09:50.539 STEP: Waiting for a default service account to be provisioned in namespace 11/26/22 05:09:50.702 STEP: Waiting for kube-root-ca.crt to be provisioned in namespace 11/26/22 05:09:50.801 [BeforeEach] [sig-network] LoadBalancers ESIPP [Slow] test/e2e/framework/metrics/init/init.go:31 [BeforeEach] [sig-network] LoadBalancers ESIPP [Slow] test/e2e/network/loadbalancer.go:1250 [It] should handle updates to ExternalTrafficPolicy field test/e2e/network/loadbalancer.go:1480 STEP: creating a service esipp-2673/external-local-update with type=LoadBalancer 11/26/22 05:09:51.075 STEP: setting ExternalTrafficPolicy=Local 11/26/22 05:09:51.075 STEP: waiting for loadbalancer for service esipp-2673/external-local-update 11/26/22 05:09:51.208 Nov 26 05:09:51.208: INFO: Waiting up to 15m0s for service "external-local-update" to have a LoadBalancer STEP: creating a pod to be part of the service external-local-update 11/26/22 05:11:23.42 Nov 26 05:11:23.541: INFO: Waiting up to 2m0s for 1 pods to be created Nov 26 05:11:23.611: INFO: Found 0/1 pods - will retry Nov 26 05:11:25.663: INFO: Found all 1 pods Nov 26 05:11:25.663: INFO: Waiting up to 2m0s for 1 pods to be running and ready: [external-local-update-5xshb] Nov 26 05:11:25.663: INFO: Waiting up to 2m0s for pod "external-local-update-5xshb" in namespace "esipp-2673" to be "running and ready" Nov 26 05:11:25.711: INFO: Pod "external-local-update-5xshb": Phase="Pending", Reason="", readiness=false. Elapsed: 48.476388ms Nov 26 05:11:25.711: INFO: Error evaluating pod condition running and ready: want pod 'external-local-update-5xshb' on 'bootstrap-e2e-minion-group-0975' to be 'Running' but was 'Pending' Nov 26 05:11:27.767: INFO: Pod "external-local-update-5xshb": Phase="Running", Reason="", readiness=true. Elapsed: 2.104246457s Nov 26 05:11:27.767: INFO: Pod "external-local-update-5xshb" satisfied condition "running and ready" Nov 26 05:11:27.767: INFO: Wanted all 1 pods to be running and ready. Result: true. Pods: [external-local-update-5xshb] STEP: waiting for loadbalancer for service esipp-2673/external-local-update 11/26/22 05:11:27.767 Nov 26 05:11:27.767: INFO: Waiting up to 15m0s for service "external-local-update" to have a LoadBalancer STEP: turning ESIPP off 11/26/22 05:11:27.839 STEP: Performing setup for networking test in namespace esipp-2673 11/26/22 05:11:29.116 STEP: creating a selector 11/26/22 05:11:29.116 STEP: Creating the service pods in kubernetes 11/26/22 05:11:29.116 Nov 26 05:11:29.116: INFO: Waiting up to 10m0s for all (but 0) nodes to be schedulable Nov 26 05:11:29.460: INFO: Waiting up to 5m0s for pod "netserver-0" in namespace "esipp-2673" to be "running and ready" Nov 26 05:11:29.569: INFO: Pod "netserver-0": Phase="Pending", Reason="", readiness=false. Elapsed: 108.989371ms Nov 26 05:11:29.569: INFO: The phase of Pod netserver-0 is Pending, waiting for it to be Running (with Ready = true) Nov 26 05:11:31.619: INFO: Pod "netserver-0": Phase="Running", Reason="", readiness=false. Elapsed: 2.159720974s Nov 26 05:11:31.619: INFO: The phase of Pod netserver-0 is Running (Ready = false) Nov 26 05:11:33.645: INFO: Pod "netserver-0": Phase="Running", Reason="", readiness=false. Elapsed: 4.184989134s Nov 26 05:11:33.645: INFO: The phase of Pod netserver-0 is Running (Ready = false) Nov 26 05:11:35.621: INFO: Pod "netserver-0": Phase="Running", Reason="", readiness=false. Elapsed: 6.161557516s Nov 26 05:11:35.621: INFO: The phase of Pod netserver-0 is Running (Ready = false) Nov 26 05:11:37.647: INFO: Pod "netserver-0": Phase="Running", Reason="", readiness=false. Elapsed: 8.187795337s Nov 26 05:11:37.647: INFO: The phase of Pod netserver-0 is Running (Ready = false) Nov 26 05:11:39.621: INFO: Pod "netserver-0": Phase="Running", Reason="", readiness=false. Elapsed: 10.161520414s Nov 26 05:11:39.621: INFO: The phase of Pod netserver-0 is Running (Ready = false) Nov 26 05:11:41.627: INFO: Pod "netserver-0": Phase="Running", Reason="", readiness=false. Elapsed: 12.16774701s Nov 26 05:11:41.627: INFO: The phase of Pod netserver-0 is Running (Ready = false) Nov 26 05:11:43.672: INFO: Pod "netserver-0": Phase="Running", Reason="", readiness=false. Elapsed: 14.212853915s Nov 26 05:11:43.672: INFO: The phase of Pod netserver-0 is Running (Ready = false) Nov 26 05:11:45.675: INFO: Pod "netserver-0": Phase="Running", Reason="", readiness=false. Elapsed: 16.215488841s Nov 26 05:11:45.675: INFO: The phase of Pod netserver-0 is Running (Ready = false) Nov 26 05:11:47.639: INFO: Pod "netserver-0": Phase="Running", Reason="", readiness=false. Elapsed: 18.178975878s Nov 26 05:11:47.639: INFO: The phase of Pod netserver-0 is Running (Ready = false) Nov 26 05:11:49.642: INFO: Pod "netserver-0": Phase="Running", Reason="", readiness=false. Elapsed: 20.182134524s Nov 26 05:11:49.642: INFO: The phase of Pod netserver-0 is Running (Ready = false) Nov 26 05:11:51.616: INFO: Pod "netserver-0": Phase="Running", Reason="", readiness=true. Elapsed: 22.156426402s Nov 26 05:11:51.616: INFO: The phase of Pod netserver-0 is Running (Ready = true) Nov 26 05:11:51.616: INFO: Pod "netserver-0" satisfied condition "running and ready" Nov 26 05:11:51.679: INFO: Waiting up to 5m0s for pod "netserver-1" in namespace "esipp-2673" to be "running and ready" Nov 26 05:11:51.741: INFO: Pod "netserver-1": Phase="Running", Reason="", readiness=false. Elapsed: 62.406024ms Nov 26 05:11:51.741: INFO: The phase of Pod netserver-1 is Running (Ready = false) Nov 26 05:11:53.794: INFO: Pod "netserver-1": Phase="Running", Reason="", readiness=false. Elapsed: 2.115694673s Nov 26 05:11:53.794: INFO: The phase of Pod netserver-1 is Running (Ready = false) Nov 26 05:11:55.855: INFO: Pod "netserver-1": Phase="Running", Reason="", readiness=false. Elapsed: 4.176287826s Nov 26 05:11:55.855: INFO: The phase of Pod netserver-1 is Running (Ready = false) Nov 26 05:11:57.802: INFO: Pod "netserver-1": Phase="Running", Reason="", readiness=false. Elapsed: 6.122939538s Nov 26 05:11:57.802: INFO: The phase of Pod netserver-1 is Running (Ready = false) Nov 26 05:11:59.809: INFO: Pod "netserver-1": Phase="Running", Reason="", readiness=false. Elapsed: 8.130570726s Nov 26 05:11:59.809: INFO: The phase of Pod netserver-1 is Running (Ready = false) Nov 26 05:12:01.798: INFO: Pod "netserver-1": Phase="Running", Reason="", readiness=false. Elapsed: 10.119558532s Nov 26 05:12:01.798: INFO: The phase of Pod netserver-1 is Running (Ready = false) Nov 26 05:12:03.817: INFO: Pod "netserver-1": Phase="Running", Reason="", readiness=false. Elapsed: 12.138647177s Nov 26 05:12:03.817: INFO: The phase of Pod netserver-1 is Running (Ready = false) Nov 26 05:12:05.797: INFO: Pod "netserver-1": Phase="Running", Reason="", readiness=false. Elapsed: 14.118828246s Nov 26 05:12:05.798: INFO: The phase of Pod netserver-1 is Running (Ready = false) Nov 26 05:12:07.939: INFO: Pod "netserver-1": Phase="Running", Reason="", readiness=false. Elapsed: 16.260606165s Nov 26 05:12:07.939: INFO: The phase of Pod netserver-1 is Running (Ready = false) Nov 26 05:12:09.795: INFO: Pod "netserver-1": Phase="Running", Reason="", readiness=false. Elapsed: 18.116033017s Nov 26 05:12:09.795: INFO: The phase of Pod netserver-1 is Running (Ready = false) Nov 26 05:12:11.790: INFO: Pod "netserver-1": Phase="Running", Reason="", readiness=false. Elapsed: 20.11176902s Nov 26 05:12:11.790: INFO: The phase of Pod netserver-1 is Running (Ready = false) Nov 26 05:12:13.806: INFO: Pod "netserver-1": Phase="Running", Reason="", readiness=false. Elapsed: 22.127149967s Nov 26 05:12:13.806: INFO: The phase of Pod netserver-1 is Running (Ready = false) Nov 26 05:12:15.794: INFO: Pod "netserver-1": Phase="Running", Reason="", readiness=false. Elapsed: 24.114918489s Nov 26 05:12:15.794: INFO: The phase of Pod netserver-1 is Running (Ready = false) Nov 26 05:12:17.799: INFO: Pod "netserver-1": Phase="Running", Reason="", readiness=false. Elapsed: 26.120351365s Nov 26 05:12:17.799: INFO: The phase of Pod netserver-1 is Running (Ready = false) Nov 26 05:12:19.806: INFO: Pod "netserver-1": Phase="Running", Reason="", readiness=false. Elapsed: 28.127288575s Nov 26 05:12:19.806: INFO: The phase of Pod netserver-1 is Running (Ready = false) Nov 26 05:12:21.792: INFO: Pod "netserver-1": Phase="Running", Reason="", readiness=false. Elapsed: 30.113196328s Nov 26 05:12:21.792: INFO: The phase of Pod netserver-1 is Running (Ready = false) Nov 26 05:12:23.803: INFO: Pod "netserver-1": Phase="Running", Reason="", readiness=false. Elapsed: 32.123961933s Nov 26 05:12:23.803: INFO: The phase of Pod netserver-1 is Running (Ready = false) Nov 26 05:12:25.791: INFO: Pod "netserver-1": Phase="Running", Reason="", readiness=false. Elapsed: 34.112387657s Nov 26 05:12:25.791: INFO: The phase of Pod netserver-1 is Running (Ready = false) Nov 26 05:12:27.812: INFO: Pod "netserver-1": Phase="Running", Reason="", readiness=false. Elapsed: 36.132862647s Nov 26 05:12:27.812: INFO: The phase of Pod netserver-1 is Running (Ready = false) Nov 26 05:12:29.797: INFO: Pod "netserver-1": Phase="Running", Reason="", readiness=false. Elapsed: 38.118563669s Nov 26 05:12:29.797: INFO: The phase of Pod netserver-1 is Running (Ready = false) Nov 26 05:12:31.815: INFO: Pod "netserver-1": Phase="Running", Reason="", readiness=false. Elapsed: 40.136703124s Nov 26 05:12:31.815: INFO: The phase of Pod netserver-1 is Running (Ready = false) Nov 26 05:12:33.810: INFO: Pod "netserver-1": Phase="Running", Reason="", readiness=false. Elapsed: 42.131571975s Nov 26 05:12:33.810: INFO: The phase of Pod netserver-1 is Running (Ready = false) Nov 26 05:12:35.787: INFO: Pod "netserver-1": Phase="Running", Reason="", readiness=false. Elapsed: 44.108565427s Nov 26 05:12:35.787: INFO: The phase of Pod netserver-1 is Running (Ready = false) Nov 26 05:12:37.818: INFO: Pod "netserver-1": Phase="Running", Reason="", readiness=false. Elapsed: 46.139311401s Nov 26 05:12:37.818: INFO: The phase of Pod netserver-1 is Running (Ready = false) Nov 26 05:12:39.794: INFO: Pod "netserver-1": Phase="Running", Reason="", readiness=false. Elapsed: 48.114844111s Nov 26 05:12:39.794: INFO: The phase of Pod netserver-1 is Running (Ready = false) Nov 26 05:12:41.791: INFO: Pod "netserver-1": Phase="Running", Reason="", readiness=false. Elapsed: 50.112101463s Nov 26 05:12:41.791: INFO: The phase of Pod netserver-1 is Running (Ready = false) Nov 26 05:12:43.797: INFO: Pod "netserver-1": Phase="Running", Reason="", readiness=false. Elapsed: 52.117928345s Nov 26 05:12:43.797: INFO: The phase of Pod netserver-1 is Running (Ready = false) Nov 26 05:12:45.792: INFO: Pod "netserver-1": Phase="Running", Reason="", readiness=false. Elapsed: 54.113825544s Nov 26 05:12:45.793: INFO: The phase of Pod netserver-1 is Running (Ready = false) Nov 26 05:12:47.804: INFO: Pod "netserver-1": Phase="Running", Reason="", readiness=false. Elapsed: 56.125287304s Nov 26 05:12:47.804: INFO: The phase of Pod netserver-1 is Running (Ready = false) Nov 26 05:12:49.832: INFO: Pod "netserver-1": Phase="Running", Reason="", readiness=false. Elapsed: 58.152892731s Nov 26 05:12:49.832: INFO: The phase of Pod netserver-1 is Running (Ready = false) Nov 26 05:12:51.797: INFO: Pod "netserver-1": Phase="Running", Reason="", readiness=false. Elapsed: 1m0.118320357s Nov 26 05:12:51.797: INFO: The phase of Pod netserver-1 is Running (Ready = false) Nov 26 05:12:53.793: INFO: Pod "netserver-1": Phase="Running", Reason="", readiness=false. Elapsed: 1m2.114821535s Nov 26 05:12:53.794: INFO: The phase of Pod netserver-1 is Running (Ready = false) Nov 26 05:12:55.803: INFO: Pod "netserver-1": Phase="Running", Reason="", readiness=false. Elapsed: 1m4.124808988s Nov 26 05:12:55.803: INFO: The phase of Pod netserver-1 is Running (Ready = false) Nov 26 05:12:57.790: INFO: Pod "netserver-1": Phase="Running", Reason="", readiness=false. Elapsed: 1m6.111277266s Nov 26 05:12:57.790: INFO: The phase of Pod netserver-1 is Running (Ready = false) Nov 26 05:12:59.800: INFO: Pod "netserver-1": Phase="Running", Reason="", readiness=false. Elapsed: 1m8.121192012s Nov 26 05:12:59.800: INFO: The phase of Pod netserver-1 is Running (Ready = false) Nov 26 05:13:01.798: INFO: Pod "netserver-1": Phase="Running", Reason="", readiness=false. Elapsed: 1m10.119032168s Nov 26 05:13:01.798: INFO: The phase of Pod netserver-1 is Running (Ready = false) Nov 26 05:13:03.820: INFO: Pod "netserver-1": Phase="Running", Reason="", readiness=false. Elapsed: 1m12.14103316s Nov 26 05:13:03.820: INFO: The phase of Pod netserver-1 is Running (Ready = false) Nov 26 05:13:05.819: INFO: Pod "netserver-1": Phase="Running", Reason="", readiness=false. Elapsed: 1m14.140188804s Nov 26 05:13:05.819: INFO: The phase of Pod netserver-1 is Running (Ready = false) Nov 26 05:13:07.815: INFO: Pod "netserver-1": Phase="Running", Reason="", readiness=false. Elapsed: 1m16.136777934s Nov 26 05:13:07.815: INFO: The phase of Pod netserver-1 is Running (Ready = false) Nov 26 05:13:09.860: INFO: Pod "netserver-1": Phase="Running", Reason="", readiness=false. Elapsed: 1m18.181791775s Nov 26 05:13:09.860: INFO: The phase of Pod netserver-1 is Running (Ready = false) Nov 26 05:13:11.790: INFO: Pod "netserver-1": Phase="Running", Reason="", readiness=false. Elapsed: 1m20.110990115s Nov 26 05:13:11.790: INFO: The phase of Pod netserver-1 is Running (Ready = false) Nov 26 05:13:13.860: INFO: Pod "netserver-1": Phase="Running", Reason="", readiness=false. Elapsed: 1m22.181020941s Nov 26 05:13:13.860: INFO: The phase of Pod netserver-1 is Running (Ready = false) Nov 26 05:13:15.806: INFO: Pod "netserver-1": Phase="Running", Reason="", readiness=false. Elapsed: 1m24.127603314s Nov 26 05:13:15.806: INFO: The phase of Pod netserver-1 is Running (Ready = false) Nov 26 05:13:17.801: INFO: Pod "netserver-1": Phase="Running", Reason="", readiness=false. Elapsed: 1m26.1224932s Nov 26 05:13:17.801: INFO: The phase of Pod netserver-1 is Running (Ready = false) Nov 26 05:13:19.793: INFO: Pod "netserver-1": Phase="Running", Reason="", readiness=false. Elapsed: 1m28.114311688s Nov 26 05:13:19.793: INFO: The phase of Pod netserver-1 is Running (Ready = false) Nov 26 05:13:21.813: INFO: Pod "netserver-1": Phase="Running", Reason="", readiness=false. Elapsed: 1m30.13480508s Nov 26 05:13:21.813: INFO: The phase of Pod netserver-1 is Running (Ready = false) Nov 26 05:13:23.812: INFO: Pod "netserver-1": Phase="Running", Reason="", readiness=false. Elapsed: 1m32.133814169s Nov 26 05:13:23.813: INFO: The phase of Pod netserver-1 is Running (Ready = false) Nov 26 05:13:25.795: INFO: Pod "netserver-1": Phase="Running", Reason="", readiness=false. Elapsed: 1m34.116658171s Nov 26 05:13:25.795: INFO: The phase of Pod netserver-1 is Running (Ready = false) Nov 26 05:13:27.822: INFO: Pod "netserver-1": Phase="Running", Reason="", readiness=false. Elapsed: 1m36.142881757s Nov 26 05:13:27.822: INFO: The phase of Pod netserver-1 is Running (Ready = false) Nov 26 05:13:29.792: INFO: Pod "netserver-1": Phase="Running", Reason="", readiness=false. Elapsed: 1m38.112905143s Nov 26 05:13:29.792: INFO: The phase of Pod netserver-1 is Running (Ready = false) Nov 26 05:13:31.799: INFO: Pod "netserver-1": Phase="Running", Reason="", readiness=false. Elapsed: 1m40.120395847s Nov 26 05:13:31.799: INFO: The phase of Pod netserver-1 is Running (Ready = false) Nov 26 05:13:33.808: INFO: Pod "netserver-1": Phase="Running", Reason="", readiness=false. Elapsed: 1m42.12921916s Nov 26 05:13:33.808: INFO: The phase of Pod netserver-1 is Running (Ready = false) Nov 26 05:13:35.824: INFO: Pod "netserver-1": Phase="Running", Reason="", readiness=false. Elapsed: 1m44.145050072s Nov 26 05:13:35.824: INFO: The phase of Pod netserver-1 is Running (Ready = false) Nov 26 05:13:37.793: INFO: Pod "netserver-1": Phase="Running", Reason="", readiness=false. Elapsed: 1m46.114416976s Nov 26 05:13:37.793: INFO: The phase of Pod netserver-1 is Running (Ready = false) Nov 26 05:13:39.820: INFO: Pod "netserver-1": Phase="Running", Reason="", readiness=false. Elapsed: 1m48.141208985s Nov 26 05:13:39.820: INFO: The phase of Pod netserver-1 is Running (Ready = false) Nov 26 05:13:41.794: INFO: Pod "netserver-1": Phase="Running", Reason="", readiness=false. Elapsed: 1m50.115487829s Nov 26 05:13:41.794: INFO: The phase of Pod netserver-1 is Running (Ready = false) Nov 26 05:13:43.802: INFO: Pod "netserver-1": Phase="Running", Reason="", readiness=false. Elapsed: 1m52.123627727s Nov 26 05:13:43.802: INFO: The phase of Pod netserver-1 is Running (Ready = false) Nov 26 05:13:45.788: INFO: Pod "netserver-1": Phase="Running", Reason="", readiness=false. Elapsed: 1m54.109503601s Nov 26 05:13:45.788: INFO: The phase of Pod netserver-1 is Running (Ready = false) Nov 26 05:13:47.791: INFO: Pod "netserver-1": Phase="Running", Reason="", readiness=false. Elapsed: 1m56.112612361s Nov 26 05:13:47.791: INFO: The phase of Pod netserver-1 is Running (Ready = false) Nov 26 05:13:49.794: INFO: Pod "netserver-1": Phase="Running", Reason="", readiness=false. Elapsed: 1m58.115716623s Nov 26 05:13:49.794: INFO: The phase of Pod netserver-1 is Running (Ready = false) Nov 26 05:13:51.786: INFO: Pod "netserver-1": Phase="Running", Reason="", readiness=false. Elapsed: 2m0.107500771s Nov 26 05:13:51.786: INFO: The phase of Pod netserver-1 is Running (Ready = false) Nov 26 05:13:53.787: INFO: Pod "netserver-1": Phase="Running", Reason="", readiness=false. Elapsed: 2m2.108512335s Nov 26 05:13:53.787: INFO: The phase of Pod netserver-1 is Running (Ready = false) Nov 26 05:13:55.786: INFO: Pod "netserver-1": Phase="Running", Reason="", readiness=false. Elapsed: 2m4.106875697s Nov 26 05:13:55.786: INFO: The phase of Pod netserver-1 is Running (Ready = false) Nov 26 05:13:57.781: INFO: Encountered non-retryable error while getting pod esipp-2673/netserver-1: Get "https://35.227.179.144/api/v1/namespaces/esipp-2673/pods/netserver-1": dial tcp 35.227.179.144:443: connect: connection refused Nov 26 05:13:57.781: INFO: Unexpected error: <*fmt.wrapError | 0xc0013600c0>: { msg: "error while waiting for pod esipp-2673/netserver-1 to be running and ready: Get \"https://35.227.179.144/api/v1/namespaces/esipp-2673/pods/netserver-1\": dial tcp 35.227.179.144:443: connect: connection refused", err: <*url.Error | 0xc001ffe240>{ Op: "Get", URL: "https://35.227.179.144/api/v1/namespaces/esipp-2673/pods/netserver-1", Err: <*net.OpError | 0xc00385e190>{ Op: "dial", Net: "tcp", Source: nil, Addr: <*net.TCPAddr | 0xc0021eede0>{ IP: [0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 255, 255, 35, 227, 179, 144], Port: 443, Zone: "", }, Err: <*os.SyscallError | 0xc001360080>{ Syscall: "connect", Err: <syscall.Errno>0x6f, }, }, }, } Nov 26 05:13:57.781: FAIL: error while waiting for pod esipp-2673/netserver-1 to be running and ready: Get "https://35.227.179.144/api/v1/namespaces/esipp-2673/pods/netserver-1": dial tcp 35.227.179.144:443: connect: connection refused Full Stack Trace k8s.io/kubernetes/test/e2e/framework/network.(*NetworkingTestConfig).createNetProxyPods(0xc001006b60, {0x75c6f7c, 0x9}, 0xc0027b59e0) test/e2e/framework/network/utils.go:866 +0x1d0 k8s.io/kubernetes/test/e2e/framework/network.(*NetworkingTestConfig).setupCore(0xc001006b60, 0x7faf5c713740?) test/e2e/framework/network/utils.go:763 +0x55 k8s.io/kubernetes/test/e2e/framework/network.(*NetworkingTestConfig).setup(0xc001006b60, 0x3c?) test/e2e/framework/network/utils.go:778 +0x3e k8s.io/kubernetes/test/e2e/framework/network.NewNetworkingTestConfig(0xc001228000, {0x0, 0x0, 0x7f8f6d0?}) test/e2e/framework/network/utils.go:131 +0x125 k8s.io/kubernetes/test/e2e/network.glob..func20.7() test/e2e/network/loadbalancer.go:1544 +0x417 Nov 26 05:13:57.821: INFO: Unexpected error: <*errors.errorString | 0xc00163aca0>: { s: "failed to get Service \"external-local-update\": Get \"https://35.227.179.144/api/v1/namespaces/esipp-2673/services/external-local-update\": dial tcp 35.227.179.144:443: connect: connection refused", } Nov 26 05:13:57.821: FAIL: failed to get Service "external-local-update": Get "https://35.227.179.144/api/v1/namespaces/esipp-2673/services/external-local-update": dial tcp 35.227.179.144:443: connect: connection refused Full Stack Trace k8s.io/kubernetes/test/e2e/network.glob..func20.7.1() test/e2e/network/loadbalancer.go:1495 +0xae panic({0x70eb7e0, 0xc00037e8c0}) /usr/local/go/src/runtime/panic.go:884 +0x212 k8s.io/kubernetes/test/e2e/framework.Fail({0xc000177d40, 0xd0}, {0xc00097b0e0?, 0xc000177d40?, 0xc00097b108?}) test/e2e/framework/log.go:61 +0x145 k8s.io/kubernetes/test/e2e/framework.ExpectNoErrorWithOffset(0x1, {0x7fa3f20, 0xc0013600c0}, {0x0?, 0xc0031c0ba0?, 0x0?}) test/e2e/framework/expect.go:76 +0x267 k8s.io/kubernetes/test/e2e/framework.ExpectNoError(...) test/e2e/framework/expect.go:43 k8s.io/kubernetes/test/e2e/framework/network.(*NetworkingTestConfig).createNetProxyPods(0xc001006b60, {0x75c6f7c, 0x9}, 0xc0027b59e0) test/e2e/framework/network/utils.go:866 +0x1d0 k8s.io/kubernetes/test/e2e/framework/network.(*NetworkingTestConfig).setupCore(0xc001006b60, 0x7faf5c713740?) test/e2e/framework/network/utils.go:763 +0x55 k8s.io/kubernetes/test/e2e/framework/network.(*NetworkingTestConfig).setup(0xc001006b60, 0x3c?) test/e2e/framework/network/utils.go:778 +0x3e k8s.io/kubernetes/test/e2e/framework/network.NewNetworkingTestConfig(0xc001228000, {0x0, 0x0, 0x7f8f6d0?}) test/e2e/framework/network/utils.go:131 +0x125 k8s.io/kubernetes/test/e2e/network.glob..func20.7() test/e2e/network/loadbalancer.go:1544 +0x417 [AfterEach] [sig-network] LoadBalancers ESIPP [Slow] test/e2e/framework/node/init/init.go:32 Nov 26 05:13:57.822: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready [AfterEach] [sig-network] LoadBalancers ESIPP [Slow] test/e2e/network/loadbalancer.go:1260 Nov 26 05:13:57.861: INFO: Output of kubectl describe svc: Nov 26 05:13:57.861: INFO: Running '/workspace/github.com/containerd/containerd/kubernetes/platforms/linux/amd64/kubectl --server=https://35.227.179.144 --kubeconfig=/workspace/.kube/config --namespace=esipp-2673 describe svc --namespace=esipp-2673' Nov 26 05:13:57.969: INFO: rc: 1 Nov 26 05:13:57.969: INFO: [DeferCleanup (Each)] [sig-network] LoadBalancers ESIPP [Slow] test/e2e/framework/metrics/init/init.go:33 [DeferCleanup (Each)] [sig-network] LoadBalancers ESIPP [Slow] dump namespaces | framework.go:196 STEP: dump namespace information after failure 11/26/22 05:13:57.969 STEP: Collecting events from namespace "esipp-2673". 11/26/22 05:13:57.97 Nov 26 05:13:58.009: INFO: Unexpected error: failed to list events in namespace "esipp-2673": <*url.Error | 0xc001ffea20>: { Op: "Get", URL: "https://35.227.179.144/api/v1/namespaces/esipp-2673/events", Err: <*net.OpError | 0xc00385e500>{ Op: "dial", Net: "tcp", Source: nil, Addr: <*net.TCPAddr | 0xc001e30720>{ IP: [0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 255, 255, 35, 227, 179, 144], Port: 443, Zone: "", }, Err: <*os.SyscallError | 0xc0013605a0>{ Syscall: "connect", Err: <syscall.Errno>0x6f, }, }, } Nov 26 05:13:58.009: FAIL: failed to list events in namespace "esipp-2673": Get "https://35.227.179.144/api/v1/namespaces/esipp-2673/events": dial tcp 35.227.179.144:443: connect: connection refused Full Stack Trace k8s.io/kubernetes/test/e2e/framework/debug.dumpEventsInNamespace(0xc00097a5c0, {0xc0031c0ba0, 0xa}) test/e2e/framework/debug/dump.go:44 +0x191 k8s.io/kubernetes/test/e2e/framework/debug.DumpAllNamespaceInfo({0x801de88, 0xc000af7520}, {0xc0031c0ba0, 0xa}) test/e2e/framework/debug/dump.go:62 +0x8d k8s.io/kubernetes/test/e2e/framework/debug/init.init.0.func1.1(0xc00097a650?, {0xc0031c0ba0?, 0x7fa7740?}) test/e2e/framework/debug/init/init.go:34 +0x32 k8s.io/kubernetes/test/e2e/framework.(*Framework).dumpNamespaceInfo.func1() test/e2e/framework/framework.go:274 +0x6d k8s.io/kubernetes/test/e2e/framework.(*Framework).dumpNamespaceInfo(0xc001228000) test/e2e/framework/framework.go:271 +0x179 reflect.Value.call({0x6627cc0?, 0xc001586290?, 0x0?}, {0x75b6e72, 0x4}, {0xae73300, 0x0, 0x0?}) /usr/local/go/src/reflect/value.go:584 +0x8c5 reflect.Value.Call({0x6627cc0?, 0xc001586290?, 0x0?}, {0xae73300?, 0x0?, 0x0?}) /usr/local/go/src/reflect/value.go:368 +0xbc [DeferCleanup (Each)] [sig-network] LoadBalancers ESIPP [Slow] tear down framework | framework.go:193 STEP: Destroying namespace "esipp-2673" for this suite. 11/26/22 05:13:58.009 Nov 26 05:13:58.049: FAIL: Couldn't delete ns: "esipp-2673": Delete "https://35.227.179.144/api/v1/namespaces/esipp-2673": dial tcp 35.227.179.144:443: connect: connection refused (&url.Error{Op:"Delete", URL:"https://35.227.179.144/api/v1/namespaces/esipp-2673", Err:(*net.OpError)(0xc0018e0d20)}) Full Stack Trace k8s.io/kubernetes/test/e2e/framework.(*Framework).AfterEach.func1() test/e2e/framework/framework.go:370 +0x4fe k8s.io/kubernetes/test/e2e/framework.(*Framework).AfterEach(0xc001228000) test/e2e/framework/framework.go:383 +0x1ca reflect.Value.call({0x6627cc0?, 0xc0015861e0?, 0xc00360ffb0?}, {0x75b6e72, 0x4}, {0xae73300, 0x0, 0x0?}) /usr/local/go/src/reflect/value.go:584 +0x8c5 reflect.Value.Call({0x6627cc0?, 0xc0015861e0?, 0x0?}, {0xae73300?, 0x5?, 0xc000526a68?}) /usr/local/go/src/reflect/value.go:368 +0xbc
Filter through log files | View test history on testgrid
go run hack/e2e.go -v --test --test_args='--ginkgo.focus=Kubernetes\se2e\ssuite\s\[It\]\s\[sig\-network\]\sLoadBalancers\sESIPP\s\[Slow\]\sshould\sonly\starget\snodes\swith\sendpoints$'
test/e2e/framework/network/utils.go:834 k8s.io/kubernetes/test/e2e/framework/network.(*NetworkingTestConfig).setup(0xc0043a9b20, 0x3c?) test/e2e/framework/network/utils.go:834 +0x545 k8s.io/kubernetes/test/e2e/framework/network.NewNetworkingTestConfig(0xc001242000, {0x0, 0x0, 0xc00464f6c0?}) test/e2e/framework/network/utils.go:131 +0x125 k8s.io/kubernetes/test/e2e/network.glob..func20.5() test/e2e/network/loadbalancer.go:1382 +0x445from junit_01.xml
[BeforeEach] [sig-network] LoadBalancers ESIPP [Slow] set up framework | framework.go:178 STEP: Creating a kubernetes client 11/26/22 05:09:07.259 Nov 26 05:09:07.259: INFO: >>> kubeConfig: /workspace/.kube/config STEP: Building a namespace api object, basename esipp 11/26/22 05:09:07.26 STEP: Waiting for a default service account to be provisioned in namespace 11/26/22 05:09:42.639 STEP: Waiting for kube-root-ca.crt to be provisioned in namespace 11/26/22 05:09:42.728 [BeforeEach] [sig-network] LoadBalancers ESIPP [Slow] test/e2e/framework/metrics/init/init.go:31 [BeforeEach] [sig-network] LoadBalancers ESIPP [Slow] test/e2e/network/loadbalancer.go:1250 [It] should only target nodes with endpoints test/e2e/network/loadbalancer.go:1346 STEP: creating a service esipp-2129/external-local-nodes with type=LoadBalancer 11/26/22 05:09:42.947 STEP: setting ExternalTrafficPolicy=Local 11/26/22 05:09:42.947 STEP: waiting for loadbalancer for service esipp-2129/external-local-nodes 11/26/22 05:09:43.027 Nov 26 05:09:43.027: INFO: Waiting up to 15m0s for service "external-local-nodes" to have a LoadBalancer STEP: waiting for loadbalancer for service esipp-2129/external-local-nodes 11/26/22 05:10:47.142 Nov 26 05:10:47.143: INFO: Waiting up to 15m0s for service "external-local-nodes" to have a LoadBalancer STEP: Performing setup for networking test in namespace esipp-2129 11/26/22 05:10:47.2 STEP: creating a selector 11/26/22 05:10:47.2 STEP: Creating the service pods in kubernetes 11/26/22 05:10:47.2 Nov 26 05:10:47.200: INFO: Waiting up to 10m0s for all (but 0) nodes to be schedulable Nov 26 05:10:47.528: INFO: Waiting up to 5m0s for pod "netserver-0" in namespace "esipp-2129" to be "running and ready" Nov 26 05:10:47.614: INFO: Pod "netserver-0": Phase="Pending", Reason="", readiness=false. Elapsed: 86.771795ms Nov 26 05:10:47.614: INFO: The phase of Pod netserver-0 is Pending, waiting for it to be Running (with Ready = true) Nov 26 05:10:49.692: INFO: Pod "netserver-0": Phase="Pending", Reason="", readiness=false. Elapsed: 2.164073098s Nov 26 05:10:49.692: INFO: The phase of Pod netserver-0 is Pending, waiting for it to be Running (with Ready = true) Nov 26 05:10:51.672: INFO: Pod "netserver-0": Phase="Pending", Reason="", readiness=false. Elapsed: 4.144191415s Nov 26 05:10:51.672: INFO: The phase of Pod netserver-0 is Pending, waiting for it to be Running (with Ready = true) Nov 26 05:10:53.693: INFO: Pod "netserver-0": Phase="Pending", Reason="", readiness=false. Elapsed: 6.165788937s Nov 26 05:10:53.693: INFO: The phase of Pod netserver-0 is Pending, waiting for it to be Running (with Ready = true) Nov 26 05:10:55.688: INFO: Pod "netserver-0": Phase="Running", Reason="", readiness=false. Elapsed: 8.15991034s Nov 26 05:10:55.688: INFO: The phase of Pod netserver-0 is Running (Ready = false) Nov 26 05:10:57.670: INFO: Pod "netserver-0": Phase="Running", Reason="", readiness=false. Elapsed: 10.142317737s Nov 26 05:10:57.670: INFO: The phase of Pod netserver-0 is Running (Ready = false) Nov 26 05:10:59.666: INFO: Pod "netserver-0": Phase="Running", Reason="", readiness=false. Elapsed: 12.137979105s Nov 26 05:10:59.666: INFO: The phase of Pod netserver-0 is Running (Ready = false) Nov 26 05:11:01.673: INFO: Pod "netserver-0": Phase="Running", Reason="", readiness=true. Elapsed: 14.145535828s Nov 26 05:11:01.673: INFO: The phase of Pod netserver-0 is Running (Ready = true) Nov 26 05:11:01.673: INFO: Pod "netserver-0" satisfied condition "running and ready" Nov 26 05:11:01.730: INFO: Waiting up to 5m0s for pod "netserver-1" in namespace "esipp-2129" to be "running and ready" Nov 26 05:11:01.857: INFO: Pod "netserver-1": Phase="Running", Reason="", readiness=false. Elapsed: 127.155376ms Nov 26 05:11:01.857: INFO: The phase of Pod netserver-1 is Running (Ready = false) Nov 26 05:11:03.912: INFO: Pod "netserver-1": Phase="Running", Reason="", readiness=false. Elapsed: 2.181840226s Nov 26 05:11:03.912: INFO: The phase of Pod netserver-1 is Running (Ready = false) Nov 26 05:11:05.913: INFO: Pod "netserver-1": Phase="Running", Reason="", readiness=false. Elapsed: 4.183181463s Nov 26 05:11:05.913: INFO: The phase of Pod netserver-1 is Running (Ready = false) Nov 26 05:11:07.942: INFO: Pod "netserver-1": Phase="Running", Reason="", readiness=false. Elapsed: 6.211659468s Nov 26 05:11:07.942: INFO: The phase of Pod netserver-1 is Running (Ready = false) Nov 26 05:11:09.917: INFO: Pod "netserver-1": Phase="Running", Reason="", readiness=true. Elapsed: 8.186478585s Nov 26 05:11:09.917: INFO: The phase of Pod netserver-1 is Running (Ready = true) Nov 26 05:11:09.917: INFO: Pod "netserver-1" satisfied condition "running and ready" Nov 26 05:11:09.976: INFO: Waiting up to 5m0s for pod "netserver-2" in namespace "esipp-2129" to be "running and ready" Nov 26 05:11:10.027: INFO: Pod "netserver-2": Phase="Running", Reason="", readiness=true. Elapsed: 50.801536ms Nov 26 05:11:10.027: INFO: The phase of Pod netserver-2 is Running (Ready = true) Nov 26 05:11:10.027: INFO: Pod "netserver-2" satisfied condition "running and ready" STEP: Creating test pods 11/26/22 05:11:10.074 Nov 26 05:11:10.141: INFO: Waiting up to 5m0s for pod "test-container-pod" in namespace "esipp-2129" to be "running" Nov 26 05:11:10.192: INFO: Pod "test-container-pod": Phase="Pending", Reason="", readiness=false. Elapsed: 50.683199ms Nov 26 05:11:12.296: INFO: Pod "test-container-pod": Phase="Pending", Reason="", readiness=false. Elapsed: 2.155240924s Nov 26 05:11:14.242: INFO: Pod "test-container-pod": Phase="Running", Reason="", readiness=true. Elapsed: 4.101123462s Nov 26 05:11:14.242: INFO: Pod "test-container-pod" satisfied condition "running" Nov 26 05:11:14.296: INFO: Setting MaxTries for pod polling to 39 for networking test based on endpoint count 3 STEP: Getting node addresses 11/26/22 05:11:14.296 Nov 26 05:11:14.296: INFO: Waiting up to 10m0s for all (but 0) nodes to be schedulable STEP: Creating the service on top of the pods in kubernetes 11/26/22 05:11:14.431 Nov 26 05:11:14.695: INFO: Service node-port-service in namespace esipp-2129 found. Nov 26 05:11:14.938: INFO: Service session-affinity-service in namespace esipp-2129 found. STEP: Waiting for NodePort service to expose endpoint 11/26/22 05:11:14.986 Nov 26 05:11:15.987: INFO: Waiting for amount of service:node-port-service endpoints to be 3 Nov 26 05:11:16.987: INFO: Waiting for amount of service:node-port-service endpoints to be 3 Nov 26 05:11:17.987: INFO: Waiting for amount of service:node-port-service endpoints to be 3 Nov 26 05:11:18.988: INFO: Waiting for amount of service:node-port-service endpoints to be 3 Nov 26 05:11:19.987: INFO: Waiting for amount of service:node-port-service endpoints to be 3 Nov 26 05:11:20.987: INFO: Waiting for amount of service:node-port-service endpoints to be 3 Nov 26 05:11:21.987: INFO: Waiting for amount of service:node-port-service endpoints to be 3 Nov 26 05:11:22.987: INFO: Waiting for amount of service:node-port-service endpoints to be 3 Nov 26 05:11:23.987: INFO: Waiting for amount of service:node-port-service endpoints to be 3 Nov 26 05:11:24.987: INFO: Waiting for amount of service:node-port-service endpoints to be 3 Nov 26 05:11:25.987: INFO: Waiting for amount of service:node-port-service endpoints to be 3 Nov 26 05:11:26.987: INFO: Waiting for amount of service:node-port-service endpoints to be 3 Nov 26 05:11:27.987: INFO: Waiting for amount of service:node-port-service endpoints to be 3 Nov 26 05:11:28.987: INFO: Waiting for amount of service:node-port-service endpoints to be 3 Nov 26 05:11:29.987: INFO: Waiting for amount of service:node-port-service endpoints to be 3 Nov 26 05:11:30.987: INFO: Waiting for amount of service:node-port-service endpoints to be 3 Nov 26 05:11:31.988: INFO: Waiting for amount of service:node-port-service endpoints to be 3 Nov 26 05:11:32.987: INFO: Waiting for amount of service:node-port-service endpoints to be 3 Nov 26 05:11:33.987: INFO: Waiting for amount of service:node-port-service endpoints to be 3 Nov 26 05:11:34.988: INFO: Waiting for amount of service:node-port-service endpoints to be 3 Nov 26 05:11:35.987: INFO: Waiting for amount of service:node-port-service endpoints to be 3 Nov 26 05:11:36.987: INFO: Waiting for amount of service:node-port-service endpoints to be 3 Nov 26 05:11:37.988: INFO: Waiting for amount of service:node-port-service endpoints to be 3 Nov 26 05:11:38.988: INFO: Waiting for amount of service:node-port-service endpoints to be 3 Nov 26 05:11:39.987: INFO: Waiting for amount of service:node-port-service endpoints to be 3 Nov 26 05:11:40.987: INFO: Waiting for amount of service:node-port-service endpoints to be 3 Nov 26 05:11:41.987: INFO: Waiting for amount of service:node-port-service endpoints to be 3 Nov 26 05:11:42.987: INFO: Waiting for amount of service:node-port-service endpoints to be 3 Nov 26 05:11:43.988: INFO: Waiting for amount of service:node-port-service endpoints to be 3 Nov 26 05:11:44.987: INFO: Waiting for amount of service:node-port-service endpoints to be 3 Nov 26 05:11:45.050: INFO: Waiting for amount of service:node-port-service endpoints to be 3 Nov 26 05:11:45.097: INFO: Unexpected error: failed to validate endpoints for service node-port-service in namespace: esipp-2129: <*errors.errorString | 0xc0001fda10>: { s: "timed out waiting for the condition", } Nov 26 05:11:45.097: FAIL: failed to validate endpoints for service node-port-service in namespace: esipp-2129: timed out waiting for the condition Full Stack Trace k8s.io/kubernetes/test/e2e/framework/network.(*NetworkingTestConfig).setup(0xc0043a9b20, 0x3c?) test/e2e/framework/network/utils.go:834 +0x545 k8s.io/kubernetes/test/e2e/framework/network.NewNetworkingTestConfig(0xc001242000, {0x0, 0x0, 0xc00464f6c0?}) test/e2e/framework/network/utils.go:131 +0x125 k8s.io/kubernetes/test/e2e/network.glob..func20.5() test/e2e/network/loadbalancer.go:1382 +0x445 Nov 26 05:11:45.247: INFO: Waiting up to 15m0s for service "external-local-nodes" to have no LoadBalancer [AfterEach] [sig-network] LoadBalancers ESIPP [Slow] test/e2e/framework/node/init/init.go:32 Nov 26 05:11:55.578: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready [AfterEach] [sig-network] LoadBalancers ESIPP [Slow] test/e2e/network/loadbalancer.go:1260 Nov 26 05:11:55.631: INFO: Output of kubectl describe svc: Nov 26 05:11:55.631: INFO: Running '/workspace/github.com/containerd/containerd/kubernetes/platforms/linux/amd64/kubectl --server=https://35.227.179.144 --kubeconfig=/workspace/.kube/config --namespace=esipp-2129 describe svc --namespace=esipp-2129' Nov 26 05:11:56.459: INFO: stderr: "" Nov 26 05:11:56.459: INFO: stdout: "Name: external-local-nodes\nNamespace: esipp-2129\nLabels: testid=external-local-nodes-e04bddb5-2a56-407e-ac26-0c3be3fa45b5\nAnnotations: <none>\nSelector: testid=external-local-nodes-e04bddb5-2a56-407e-ac26-0c3be3fa45b5\nType: ClusterIP\nIP Family Policy: SingleStack\nIP Families: IPv4\nIP: 10.0.34.197\nIPs: 10.0.34.197\nPort: <unset> 8081/TCP\nTargetPort: 80/TCP\nEndpoints: <none>\nSession Affinity: None\nEvents:\n Type Reason Age From Message\n ---- ------ ---- ---- -------\n Normal EnsuringLoadBalancer 104s service-controller Ensuring load balancer\n Normal EnsuredLoadBalancer 70s service-controller Ensured load balancer\n Normal Type 11s service-controller LoadBalancer -> ClusterIP\n\n\nName: node-port-service\nNamespace: esipp-2129\nLabels: <none>\nAnnotations: <none>\nSelector: selector-1fc049b7-a557-4787-bf4b-ecc54434dbb9=true\nType: NodePort\nIP Family Policy: SingleStack\nIP Families: IPv4\nIP: 10.0.55.189\nIPs: 10.0.55.189\nPort: http 80/TCP\nTargetPort: 8083/TCP\nNodePort: http 30447/TCP\nEndpoints: 10.64.0.74:8083,10.64.3.65:8083\nPort: udp 90/UDP\nTargetPort: 8081/UDP\nNodePort: udp 30593/UDP\nEndpoints: 10.64.0.74:8081,10.64.3.65:8081\nSession Affinity: None\nExternal Traffic Policy: Cluster\nEvents: <none>\n\n\nName: session-affinity-service\nNamespace: esipp-2129\nLabels: <none>\nAnnotations: <none>\nSelector: selector-1fc049b7-a557-4787-bf4b-ecc54434dbb9=true\nType: NodePort\nIP Family Policy: SingleStack\nIP Families: IPv4\nIP: 10.0.188.189\nIPs: 10.0.188.189\nPort: http 80/TCP\nTargetPort: 8083/TCP\nNodePort: http 31448/TCP\nEndpoints: 10.64.0.74:8083,10.64.3.65:8083\nPort: udp 90/UDP\nTargetPort: 8081/UDP\nNodePort: udp 31670/UDP\nEndpoints: 10.64.0.74:8081,10.64.3.65:8081\nSession Affinity: ClientIP\nExternal Traffic Policy: Cluster\nEvents: <none>\n" Nov 26 05:11:56.459: INFO: Name: external-local-nodes Namespace: esipp-2129 Labels: testid=external-local-nodes-e04bddb5-2a56-407e-ac26-0c3be3fa45b5 Annotations: <none> Selector: testid=external-local-nodes-e04bddb5-2a56-407e-ac26-0c3be3fa45b5 Type: ClusterIP IP Family Policy: SingleStack IP Families: IPv4 IP: 10.0.34.197 IPs: 10.0.34.197 Port: <unset> 8081/TCP TargetPort: 80/TCP Endpoints: <none> Session Affinity: None Events: Type Reason Age From Message ---- ------ ---- ---- ------- Normal EnsuringLoadBalancer 104s service-controller Ensuring load balancer Normal EnsuredLoadBalancer 70s service-controller Ensured load balancer Normal Type 11s service-controller LoadBalancer -> ClusterIP Name: node-port-service Namespace: esipp-2129 Labels: <none> Annotations: <none> Selector: selector-1fc049b7-a557-4787-bf4b-ecc54434dbb9=true Type: NodePort IP Family Policy: SingleStack IP Families: IPv4 IP: 10.0.55.189 IPs: 10.0.55.189 Port: http 80/TCP TargetPort: 8083/TCP NodePort: http 30447/TCP Endpoints: 10.64.0.74:8083,10.64.3.65:8083 Port: udp 90/UDP TargetPort: 8081/UDP NodePort: udp 30593/UDP Endpoints: 10.64.0.74:8081,10.64.3.65:8081 Session Affinity: None External Traffic Policy: Cluster Events: <none> Name: session-affinity-service Namespace: esipp-2129 Labels: <none> Annotations: <none> Selector: selector-1fc049b7-a557-4787-bf4b-ecc54434dbb9=true Type: NodePort IP Family Policy: SingleStack IP Families: IPv4 IP: 10.0.188.189 IPs: 10.0.188.189 Port: http 80/TCP TargetPort: 8083/TCP NodePort: http 31448/TCP Endpoints: 10.64.0.74:8083,10.64.3.65:8083 Port: udp 90/UDP TargetPort: 8081/UDP NodePort: udp 31670/UDP Endpoints: 10.64.0.74:8081,10.64.3.65:8081 Session Affinity: ClientIP External Traffic Policy: Cluster Events: <none> [DeferCleanup (Each)] [sig-network] LoadBalancers ESIPP [Slow] test/e2e/framework/metrics/init/init.go:33 [DeferCleanup (Each)] [sig-network] LoadBalancers ESIPP [Slow] dump namespaces | framework.go:196 STEP: dump namespace information after failure 11/26/22 05:11:56.459 STEP: Collecting events from namespace "esipp-2129". 11/26/22 05:11:56.459 STEP: Found 26 events. 11/26/22 05:11:56.523 Nov 26 05:11:56.523: INFO: At 0001-01-01 00:00:00 +0000 UTC - event for netserver-0: { } Scheduled: Successfully assigned esipp-2129/netserver-0 to bootstrap-e2e-minion-group-0975 Nov 26 05:11:56.523: INFO: At 0001-01-01 00:00:00 +0000 UTC - event for netserver-1: { } Scheduled: Successfully assigned esipp-2129/netserver-1 to bootstrap-e2e-minion-group-rwsn Nov 26 05:11:56.523: INFO: At 0001-01-01 00:00:00 +0000 UTC - event for netserver-2: { } Scheduled: Successfully assigned esipp-2129/netserver-2 to bootstrap-e2e-minion-group-vhdj Nov 26 05:11:56.523: INFO: At 0001-01-01 00:00:00 +0000 UTC - event for test-container-pod: { } Scheduled: Successfully assigned esipp-2129/test-container-pod to bootstrap-e2e-minion-group-0975 Nov 26 05:11:56.523: INFO: At 2022-11-26 05:10:12 +0000 UTC - event for external-local-nodes: {service-controller } EnsuringLoadBalancer: Ensuring load balancer Nov 26 05:11:56.523: INFO: At 2022-11-26 05:10:46 +0000 UTC - event for external-local-nodes: {service-controller } EnsuredLoadBalancer: Ensured load balancer Nov 26 05:11:56.523: INFO: At 2022-11-26 05:10:48 +0000 UTC - event for netserver-0: {kubelet bootstrap-e2e-minion-group-0975} Pulled: Container image "registry.k8s.io/e2e-test-images/agnhost:2.43" already present on machine Nov 26 05:11:56.523: INFO: At 2022-11-26 05:10:48 +0000 UTC - event for netserver-0: {kubelet bootstrap-e2e-minion-group-0975} Created: Created container webserver Nov 26 05:11:56.523: INFO: At 2022-11-26 05:10:48 +0000 UTC - event for netserver-0: {kubelet bootstrap-e2e-minion-group-0975} Started: Started container webserver Nov 26 05:11:56.523: INFO: At 2022-11-26 05:10:48 +0000 UTC - event for netserver-2: {kubelet bootstrap-e2e-minion-group-vhdj} Created: Created container webserver Nov 26 05:11:56.523: INFO: At 2022-11-26 05:10:48 +0000 UTC - event for netserver-2: {kubelet bootstrap-e2e-minion-group-vhdj} Started: Started container webserver Nov 26 05:11:56.523: INFO: At 2022-11-26 05:10:48 +0000 UTC - event for netserver-2: {kubelet bootstrap-e2e-minion-group-vhdj} Pulled: Container image "registry.k8s.io/e2e-test-images/agnhost:2.43" already present on machine Nov 26 05:11:56.523: INFO: At 2022-11-26 05:10:49 +0000 UTC - event for netserver-1: {kubelet bootstrap-e2e-minion-group-rwsn} Pulled: Container image "registry.k8s.io/e2e-test-images/agnhost:2.43" already present on machine Nov 26 05:11:56.523: INFO: At 2022-11-26 05:10:49 +0000 UTC - event for netserver-1: {kubelet bootstrap-e2e-minion-group-rwsn} Created: Created container webserver Nov 26 05:11:56.523: INFO: At 2022-11-26 05:10:49 +0000 UTC - event for netserver-1: {kubelet bootstrap-e2e-minion-group-rwsn} Started: Started container webserver Nov 26 05:11:56.523: INFO: At 2022-11-26 05:10:49 +0000 UTC - event for netserver-2: {kubelet bootstrap-e2e-minion-group-vhdj} Killing: Stopping container webserver Nov 26 05:11:56.523: INFO: At 2022-11-26 05:10:50 +0000 UTC - event for netserver-2: {kubelet bootstrap-e2e-minion-group-vhdj} SandboxChanged: Pod sandbox changed, it will be killed and re-created. Nov 26 05:11:56.523: INFO: At 2022-11-26 05:10:51 +0000 UTC - event for netserver-1: {kubelet bootstrap-e2e-minion-group-rwsn} Killing: Stopping container webserver Nov 26 05:11:56.523: INFO: At 2022-11-26 05:10:52 +0000 UTC - event for netserver-1: {kubelet bootstrap-e2e-minion-group-rwsn} SandboxChanged: Pod sandbox changed, it will be killed and re-created. Nov 26 05:11:56.523: INFO: At 2022-11-26 05:10:58 +0000 UTC - event for netserver-0: {kubelet bootstrap-e2e-minion-group-0975} Killing: Stopping container webserver Nov 26 05:11:56.523: INFO: At 2022-11-26 05:10:59 +0000 UTC - event for netserver-0: {kubelet bootstrap-e2e-minion-group-0975} SandboxChanged: Pod sandbox changed, it will be killed and re-created. Nov 26 05:11:56.523: INFO: At 2022-11-26 05:11:04 +0000 UTC - event for netserver-0: {kubelet bootstrap-e2e-minion-group-0975} BackOff: Back-off restarting failed container webserver in pod netserver-0_esipp-2129(ef8578ae-a8ef-4e57-8843-4c9c1a2dd007) Nov 26 05:11:56.523: INFO: At 2022-11-26 05:11:11 +0000 UTC - event for test-container-pod: {kubelet bootstrap-e2e-minion-group-0975} Pulled: Container image "registry.k8s.io/e2e-test-images/agnhost:2.43" already present on machine Nov 26 05:11:56.523: INFO: At 2022-11-26 05:11:11 +0000 UTC - event for test-container-pod: {kubelet bootstrap-e2e-minion-group-0975} Created: Created container webserver Nov 26 05:11:56.523: INFO: At 2022-11-26 05:11:11 +0000 UTC - event for test-container-pod: {kubelet bootstrap-e2e-minion-group-0975} Started: Started container webserver Nov 26 05:11:56.523: INFO: At 2022-11-26 05:11:45 +0000 UTC - event for external-local-nodes: {service-controller } Type: LoadBalancer -> ClusterIP Nov 26 05:11:56.591: INFO: POD NODE PHASE GRACE CONDITIONS Nov 26 05:11:56.591: INFO: netserver-0 bootstrap-e2e-minion-group-0975 Running [{Initialized True 0001-01-01 00:00:00 +0000 UTC 2022-11-26 05:10:47 +0000 UTC } {Ready False 0001-01-01 00:00:00 +0000 UTC 2022-11-26 05:10:59 +0000 UTC ContainersNotReady containers with unready status: [webserver]} {ContainersReady False 0001-01-01 00:00:00 +0000 UTC 2022-11-26 05:10:59 +0000 UTC ContainersNotReady containers with unready status: [webserver]} {PodScheduled True 0001-01-01 00:00:00 +0000 UTC 2022-11-26 05:10:47 +0000 UTC }] Nov 26 05:11:56.592: INFO: netserver-1 bootstrap-e2e-minion-group-rwsn Running [{Initialized True 0001-01-01 00:00:00 +0000 UTC 2022-11-26 05:10:47 +0000 UTC } {Ready True 0001-01-01 00:00:00 +0000 UTC 2022-11-26 05:11:08 +0000 UTC } {ContainersReady True 0001-01-01 00:00:00 +0000 UTC 2022-11-26 05:11:08 +0000 UTC } {PodScheduled True 0001-01-01 00:00:00 +0000 UTC 2022-11-26 05:10:47 +0000 UTC }] Nov 26 05:11:56.592: INFO: netserver-2 bootstrap-e2e-minion-group-vhdj Running [{Initialized True 0001-01-01 00:00:00 +0000 UTC 2022-11-26 05:10:47 +0000 UTC } {Ready True 0001-01-01 00:00:00 +0000 UTC 2022-11-26 05:11:08 +0000 UTC } {ContainersReady True 0001-01-01 00:00:00 +0000 UTC 2022-11-26 05:11:08 +0000 UTC } {PodScheduled True 0001-01-01 00:00:00 +0000 UTC 2022-11-26 05:10:47 +0000 UTC }] Nov 26 05:11:56.592: INFO: test-container-pod bootstrap-e2e-minion-group-0975 Running [{Initialized True 0001-01-01 00:00:00 +0000 UTC 2022-11-26 05:11:10 +0000 UTC } {Ready True 0001-01-01 00:00:00 +0000 UTC 2022-11-26 05:11:11 +0000 UTC } {ContainersReady True 0001-01-01 00:00:00 +0000 UTC 2022-11-26 05:11:11 +0000 UTC } {PodScheduled True 0001-01-01 00:00:00 +0000 UTC 2022-11-26 05:11:10 +0000 UTC }] Nov 26 05:11:56.592: INFO: Nov 26 05:11:57.059: INFO: Logging node info for node bootstrap-e2e-master Nov 26 05:11:57.130: INFO: Node Info: &Node{ObjectMeta:{bootstrap-e2e-master 9a48eb11-c964-4bf8-82c0-1c32a9703981 4609 0 2022-11-26 05:05:18 +0000 UTC <nil> <nil> map[beta.kubernetes.io/arch:amd64 beta.kubernetes.io/instance-type:n1-standard-1 beta.kubernetes.io/os:linux cloud.google.com/metadata-proxy-ready:true failure-domain.beta.kubernetes.io/region:us-west1 failure-domain.beta.kubernetes.io/zone:us-west1-b kubernetes.io/arch:amd64 kubernetes.io/hostname:bootstrap-e2e-master kubernetes.io/os:linux node.kubernetes.io/instance-type:n1-standard-1 topology.kubernetes.io/region:us-west1 topology.kubernetes.io/zone:us-west1-b] map[node.alpha.kubernetes.io/ttl:0 volumes.kubernetes.io/controller-managed-attach-detach:true] [] [] [{kubelet Update v1 2022-11-26 05:05:18 +0000 UTC FieldsV1 {"f:metadata":{"f:annotations":{".":{},"f:volumes.kubernetes.io/controller-managed-attach-detach":{}},"f:labels":{".":{},"f:beta.kubernetes.io/arch":{},"f:beta.kubernetes.io/instance-type":{},"f:beta.kubernetes.io/os":{},"f:cloud.google.com/metadata-proxy-ready":{},"f:failure-domain.beta.kubernetes.io/region":{},"f:failure-domain.beta.kubernetes.io/zone":{},"f:kubernetes.io/arch":{},"f:kubernetes.io/hostname":{},"f:kubernetes.io/os":{},"f:node.kubernetes.io/instance-type":{},"f:topology.kubernetes.io/region":{},"f:topology.kubernetes.io/zone":{}}},"f:spec":{"f:providerID":{},"f:unschedulable":{}}} } {kube-controller-manager Update v1 2022-11-26 05:05:25 +0000 UTC FieldsV1 {"f:status":{"f:conditions":{"k:{\"type\":\"NetworkUnavailable\"}":{"f:lastHeartbeatTime":{},"f:lastTransitionTime":{},"f:message":{},"f:reason":{},"f:status":{}}}}} status} {kube-controller-manager Update v1 2022-11-26 05:05:39 +0000 UTC FieldsV1 {"f:metadata":{"f:annotations":{"f:node.alpha.kubernetes.io/ttl":{}}},"f:spec":{"f:podCIDR":{},"f:podCIDRs":{".":{},"v:\"10.64.1.0/24\"":{}},"f:taints":{}}} } {kubelet Update v1 2022-11-26 05:11:14 +0000 UTC FieldsV1 {"f:status":{"f:conditions":{"k:{\"type\":\"DiskPressure\"}":{"f:lastHeartbeatTime":{}},"k:{\"type\":\"MemoryPressure\"}":{"f:lastHeartbeatTime":{}},"k:{\"type\":\"PIDPressure\"}":{"f:lastHeartbeatTime":{}},"k:{\"type\":\"Ready\"}":{"f:lastHeartbeatTime":{},"f:lastTransitionTime":{},"f:message":{},"f:reason":{},"f:status":{}}},"f:images":{}}} status}]},Spec:NodeSpec{PodCIDR:10.64.1.0/24,DoNotUseExternalID:,ProviderID:gce://k8s-jkns-e2e-gce-ci-reboot/us-west1-b/bootstrap-e2e-master,Unschedulable:true,Taints:[]Taint{Taint{Key:node-role.kubernetes.io/master,Value:,Effect:NoSchedule,TimeAdded:<nil>,},Taint{Key:node.kubernetes.io/unschedulable,Value:,Effect:NoSchedule,TimeAdded:<nil>,},},ConfigSource:nil,PodCIDRs:[10.64.1.0/24],},Status:NodeStatus{Capacity:ResourceList{attachable-volumes-gce-pd: {{127 0} {<nil>} 127 DecimalSI},cpu: {{1 0} {<nil>} 1 DecimalSI},ephemeral-storage: {{16656896000 0} {<nil>} 16266500Ki BinarySI},hugepages-1Gi: {{0 0} {<nil>} 0 DecimalSI},hugepages-2Mi: {{0 0} {<nil>} 0 DecimalSI},memory: {{3858374656 0} {<nil>} BinarySI},pods: {{110 0} {<nil>} 110 DecimalSI},},Allocatable:ResourceList{attachable-volumes-gce-pd: {{127 0} {<nil>} 127 DecimalSI},cpu: {{1 0} {<nil>} 1 DecimalSI},ephemeral-storage: {{14991206376 0} {<nil>} 14991206376 DecimalSI},hugepages-1Gi: {{0 0} {<nil>} 0 DecimalSI},hugepages-2Mi: {{0 0} {<nil>} 0 DecimalSI},memory: {{3596230656 0} {<nil>} BinarySI},pods: {{110 0} {<nil>} 110 DecimalSI},},Phase:,Conditions:[]NodeCondition{NodeCondition{Type:NetworkUnavailable,Status:False,LastHeartbeatTime:2022-11-26 05:05:25 +0000 UTC,LastTransitionTime:2022-11-26 05:05:25 +0000 UTC,Reason:RouteCreated,Message:RouteController created a route,},NodeCondition{Type:MemoryPressure,Status:False,LastHeartbeatTime:2022-11-26 05:11:14 +0000 UTC,LastTransitionTime:2022-11-26 05:05:18 +0000 UTC,Reason:KubeletHasSufficientMemory,Message:kubelet has sufficient memory available,},NodeCondition{Type:DiskPressure,Status:False,LastHeartbeatTime:2022-11-26 05:11:14 +0000 UTC,LastTransitionTime:2022-11-26 05:05:18 +0000 UTC,Reason:KubeletHasNoDiskPressure,Message:kubelet has no disk pressure,},NodeCondition{Type:PIDPressure,Status:False,LastHeartbeatTime:2022-11-26 05:11:14 +0000 UTC,LastTransitionTime:2022-11-26 05:05:18 +0000 UTC,Reason:KubeletHasSufficientPID,Message:kubelet has sufficient PID available,},NodeCondition{Type:Ready,Status:True,LastHeartbeatTime:2022-11-26 05:11:14 +0000 UTC,LastTransitionTime:2022-11-26 05:05:38 +0000 UTC,Reason:KubeletReady,Message:kubelet is posting ready status. AppArmor enabled,},},Addresses:[]NodeAddress{NodeAddress{Type:InternalIP,Address:10.138.0.2,},NodeAddress{Type:ExternalIP,Address:35.227.179.144,},NodeAddress{Type:InternalDNS,Address:bootstrap-e2e-master.c.k8s-jkns-e2e-gce-ci-reboot.internal,},NodeAddress{Type:Hostname,Address:bootstrap-e2e-master.c.k8s-jkns-e2e-gce-ci-reboot.internal,},},DaemonEndpoints:NodeDaemonEndpoints{KubeletEndpoint:DaemonEndpoint{Port:10250,},},NodeInfo:NodeSystemInfo{MachineID:0fdb3cfe29f66637553465718381a2f8,SystemUUID:0fdb3cfe-29f6-6637-5534-65718381a2f8,BootID:fb804a29-7e61-434d-8044-6944159ce35f,KernelVersion:5.10.123+,OSImage:Container-Optimized OS from Google,ContainerRuntimeVersion:containerd://1.7.0-beta.0-149-gd06318622,KubeletVersion:v1.27.0-alpha.0.50+70617042976dc1,KubeProxyVersion:v1.27.0-alpha.0.50+70617042976dc1,OperatingSystem:linux,Architecture:amd64,},Images:[]ContainerImage{ContainerImage{Names:[registry.k8s.io/kube-apiserver-amd64:v1.27.0-alpha.0.50_70617042976dc1],SizeBytes:135160272,},ContainerImage{Names:[registry.k8s.io/kube-controller-manager-amd64:v1.27.0-alpha.0.50_70617042976dc1],SizeBytes:124990265,},ContainerImage{Names:[registry.k8s.io/etcd@sha256:dd75ec974b0a2a6f6bb47001ba09207976e625db898d1b16735528c009cb171c registry.k8s.io/etcd:3.5.6-0],SizeBytes:102542580,},ContainerImage{Names:[registry.k8s.io/kube-scheduler-amd64:v1.27.0-alpha.0.50_70617042976dc1],SizeBytes:57660216,},ContainerImage{Names:[gke.gcr.io/prometheus-to-sd@sha256:e739643c3939ba0b161425f45a1989eedfc4a3b166db9a7100863296b4c70510 gke.gcr.io/prometheus-to-sd:v0.11.1-gke.1],SizeBytes:48742566,},ContainerImage{Names:[gcr.io/k8s-ingress-image-push/ingress-gce-glbc-amd64@sha256:5db27383add6d9f4ebdf0286409ac31f7f5d273690204b341a4e37998917693b gcr.io/k8s-ingress-image-push/ingress-gce-glbc-amd64:v1.20.1],SizeBytes:36598135,},ContainerImage{Names:[registry.k8s.io/addon-manager/kube-addon-manager@sha256:49cc4e6e4a3745b427ce14b0141476ab339bb65c6bc05033019e046c8727dcb0 registry.k8s.io/addon-manager/kube-addon-manager:v9.1.6],SizeBytes:30464183,},ContainerImage{Names:[registry.k8s.io/kas-network-proxy/proxy-server@sha256:2c111f004bec24888d8cfa2a812a38fb8341350abac67dcd0ac64e709dfe389c registry.k8s.io/kas-network-proxy/proxy-server:v0.0.33],SizeBytes:22020129,},ContainerImage{Names:[registry.k8s.io/metadata-proxy@sha256:e914645f22e946bce5165737e1b244e0a296ad1f0f81a9531adc57af2780978a registry.k8s.io/metadata-proxy:v0.1.12],SizeBytes:5301657,},ContainerImage{Names:[registry.k8s.io/pause@sha256:9001185023633d17a2f98ff69b6ff2615b8ea02a825adffa40422f51dfdcde9d registry.k8s.io/pause:3.8],SizeBytes:311286,},},VolumesInUse:[],VolumesAttached:[]AttachedVolume{},Config:nil,},} Nov 26 05:11:57.130: INFO: Logging kubelet events for node bootstrap-e2e-master Nov 26 05:11:57.375: INFO: Logging pods the kubelet thinks is on node bootstrap-e2e-master Nov 26 05:11:57.474: INFO: etcd-server-bootstrap-e2e-master started at 2022-11-26 05:04:35 +0000 UTC (0+1 container statuses recorded) Nov 26 05:11:57.474: INFO: Container etcd-container ready: true, restart count 2 Nov 26 05:11:57.474: INFO: konnectivity-server-bootstrap-e2e-master started at 2022-11-26 05:04:35 +0000 UTC (0+1 container statuses recorded) Nov 26 05:11:57.474: INFO: Container konnectivity-server-container ready: true, restart count 0 Nov 26 05:11:57.475: INFO: kube-apiserver-bootstrap-e2e-master started at 2022-11-26 05:04:35 +0000 UTC (0+1 container statuses recorded) Nov 26 05:11:57.475: INFO: Container kube-apiserver ready: true, restart count 0 Nov 26 05:11:57.475: INFO: l7-lb-controller-bootstrap-e2e-master started at 2022-11-26 05:04:52 +0000 UTC (0+1 container statuses recorded) Nov 26 05:11:57.475: INFO: Container l7-lb-controller ready: false, restart count 4 Nov 26 05:11:57.475: INFO: metadata-proxy-v0.1-8ll2g started at 2022-11-26 05:05:51 +0000 UTC (0+2 container statuses recorded) Nov 26 05:11:57.475: INFO: Container metadata-proxy ready: true, restart count 0 Nov 26 05:11:57.475: INFO: Container prometheus-to-sd-exporter ready: true, restart count 0 Nov 26 05:11:57.475: INFO: etcd-server-events-bootstrap-e2e-master started at 2022-11-26 05:04:35 +0000 UTC (0+1 container statuses recorded) Nov 26 05:11:57.475: INFO: Container etcd-container ready: true, restart count 2 Nov 26 05:11:57.475: INFO: kube-controller-manager-bootstrap-e2e-master started at 2022-11-26 05:04:35 +0000 UTC (0+1 container statuses recorded) Nov 26 05:11:57.475: INFO: Container kube-controller-manager ready: true, restart count 2 Nov 26 05:11:57.475: INFO: kube-scheduler-bootstrap-e2e-master started at 2022-11-26 05:04:35 +0000 UTC (0+1 container statuses recorded) Nov 26 05:11:57.475: INFO: Container kube-scheduler ready: false, restart count 2 Nov 26 05:11:57.475: INFO: kube-addon-manager-bootstrap-e2e-master started at 2022-11-26 05:04:52 +0000 UTC (0+1 container statuses recorded) Nov 26 05:11:57.475: INFO: Container kube-addon-manager ready: true, restart count 1 Nov 26 05:11:57.778: INFO: Latency metrics for node bootstrap-e2e-master Nov 26 05:11:57.778: INFO: Logging node info for node bootstrap-e2e-minion-group-0975 Nov 26 05:11:57.832: INFO: Node Info: &Node{ObjectMeta:{bootstrap-e2e-minion-group-0975 6a0671d4-0e9d-4ac8-b359-65111262cfad 4226 0 2022-11-26 05:05:22 +0000 UTC <nil> <nil> map[beta.kubernetes.io/arch:amd64 beta.kubernetes.io/instance-type:n1-standard-2 beta.kubernetes.io/os:linux cloud.google.com/metadata-proxy-ready:true failure-domain.beta.kubernetes.io/region:us-west1 failure-domain.beta.kubernetes.io/zone:us-west1-b kubernetes.io/arch:amd64 kubernetes.io/hostname:bootstrap-e2e-minion-group-0975 kubernetes.io/os:linux node.kubernetes.io/instance-type:n1-standard-2 topology.hostpath.csi/node:bootstrap-e2e-minion-group-0975 topology.kubernetes.io/region:us-west1 topology.kubernetes.io/zone:us-west1-b] map[csi.volume.kubernetes.io/nodeid:{"csi-hostpath-multivolume-593":"bootstrap-e2e-minion-group-0975"} node.alpha.kubernetes.io/ttl:0 volumes.kubernetes.io/controller-managed-attach-detach:true] [] [] [{kubelet Update v1 2022-11-26 05:05:22 +0000 UTC FieldsV1 {"f:metadata":{"f:annotations":{".":{},"f:volumes.kubernetes.io/controller-managed-attach-detach":{}},"f:labels":{".":{},"f:beta.kubernetes.io/arch":{},"f:beta.kubernetes.io/instance-type":{},"f:beta.kubernetes.io/os":{},"f:cloud.google.com/metadata-proxy-ready":{},"f:failure-domain.beta.kubernetes.io/region":{},"f:failure-domain.beta.kubernetes.io/zone":{},"f:kubernetes.io/arch":{},"f:kubernetes.io/hostname":{},"f:kubernetes.io/os":{},"f:node.kubernetes.io/instance-type":{},"f:topology.kubernetes.io/region":{},"f:topology.kubernetes.io/zone":{}}},"f:spec":{"f:providerID":{}}} } {kube-controller-manager Update v1 2022-11-26 05:05:23 +0000 UTC FieldsV1 {"f:metadata":{"f:annotations":{"f:node.alpha.kubernetes.io/ttl":{}}},"f:spec":{"f:podCIDR":{},"f:podCIDRs":{".":{},"v:\"10.64.2.0/24\"":{}}}} } {node-problem-detector Update v1 2022-11-26 05:10:29 +0000 UTC FieldsV1 {"f:status":{"f:conditions":{"k:{\"type\":\"CorruptDockerOverlay2\"}":{".":{},"f:lastHeartbeatTime":{},"f:lastTransitionTime":{},"f:message":{},"f:reason":{},"f:status":{},"f:type":{}},"k:{\"type\":\"FrequentContainerdRestart\"}":{".":{},"f:lastHeartbeatTime":{},"f:lastTransitionTime":{},"f:message":{},"f:reason":{},"f:status":{},"f:type":{}},"k:{\"type\":\"FrequentDockerRestart\"}":{".":{},"f:lastHeartbeatTime":{},"f:lastTransitionTime":{},"f:message":{},"f:reason":{},"f:status":{},"f:type":{}},"k:{\"type\":\"FrequentKubeletRestart\"}":{".":{},"f:lastHeartbeatTime":{},"f:lastTransitionTime":{},"f:message":{},"f:reason":{},"f:status":{},"f:type":{}},"k:{\"type\":\"FrequentUnregisterNetDevice\"}":{".":{},"f:lastHeartbeatTime":{},"f:lastTransitionTime":{},"f:message":{},"f:reason":{},"f:status":{},"f:type":{}},"k:{\"type\":\"KernelDeadlock\"}":{".":{},"f:lastHeartbeatTime":{},"f:lastTransitionTime":{},"f:message":{},"f:reason":{},"f:status":{},"f:type":{}},"k:{\"type\":\"ReadonlyFilesystem\"}":{".":{},"f:lastHeartbeatTime":{},"f:lastTransitionTime":{},"f:message":{},"f:reason":{},"f:status":{},"f:type":{}}}}} status} {kubelet Update v1 2022-11-26 05:10:46 +0000 UTC FieldsV1 {"f:metadata":{"f:annotations":{"f:csi.volume.kubernetes.io/nodeid":{}},"f:labels":{"f:topology.hostpath.csi/node":{}}},"f:status":{"f:conditions":{"k:{\"type\":\"DiskPressure\"}":{"f:lastHeartbeatTime":{}},"k:{\"type\":\"MemoryPressure\"}":{"f:lastHeartbeatTime":{}},"k:{\"type\":\"PIDPressure\"}":{"f:lastHeartbeatTime":{}},"k:{\"type\":\"Ready\"}":{"f:lastHeartbeatTime":{},"f:lastTransitionTime":{},"f:message":{},"f:reason":{},"f:status":{}}},"f:images":{}}} status} {kube-controller-manager Update v1 2022-11-26 05:10:49 +0000 UTC FieldsV1 {"f:status":{"f:conditions":{"k:{\"type\":\"NetworkUnavailable\"}":{"f:lastHeartbeatTime":{},"f:lastTransitionTime":{},"f:message":{},"f:reason":{},"f:status":{}}}}} status}]},Spec:NodeSpec{PodCIDR:10.64.2.0/24,DoNotUseExternalID:,ProviderID:gce://k8s-jkns-e2e-gce-ci-reboot/us-west1-b/bootstrap-e2e-minion-group-0975,Unschedulable:false,Taints:[]Taint{},ConfigSource:nil,PodCIDRs:[10.64.2.0/24],},Status:NodeStatus{Capacity:ResourceList{attachable-volumes-gce-pd: {{127 0} {<nil>} 127 DecimalSI},cpu: {{2 0} {<nil>} 2 DecimalSI},ephemeral-storage: {{101203873792 0} {<nil>} 98831908Ki BinarySI},hugepages-1Gi: {{0 0} {<nil>} 0 DecimalSI},hugepages-2Mi: {{0 0} {<nil>} 0 DecimalSI},memory: {{7815430144 0} {<nil>} BinarySI},pods: {{110 0} {<nil>} 110 DecimalSI},},Allocatable:ResourceList{attachable-volumes-gce-pd: {{127 0} {<nil>} 127 DecimalSI},cpu: {{2 0} {<nil>} 2 DecimalSI},ephemeral-storage: {{91083486262 0} {<nil>} 91083486262 DecimalSI},hugepages-1Gi: {{0 0} {<nil>} 0 DecimalSI},hugepages-2Mi: {{0 0} {<nil>} 0 DecimalSI},memory: {{7553286144 0} {<nil>} BinarySI},pods: {{110 0} {<nil>} 110 DecimalSI},},Phase:,Conditions:[]NodeCondition{NodeCondition{Type:FrequentKubeletRestart,Status:False,LastHeartbeatTime:2022-11-26 05:10:29 +0000 UTC,LastTransitionTime:2022-11-26 05:05:27 +0000 UTC,Reason:NoFrequentKubeletRestart,Message:kubelet is functioning properly,},NodeCondition{Type:FrequentDockerRestart,Status:False,LastHeartbeatTime:2022-11-26 05:10:29 +0000 UTC,LastTransitionTime:2022-11-26 05:05:27 +0000 UTC,Reason:NoFrequentDockerRestart,Message:docker is functioning properly,},NodeCondition{Type:FrequentContainerdRestart,Status:False,LastHeartbeatTime:2022-11-26 05:10:29 +0000 UTC,LastTransitionTime:2022-11-26 05:05:27 +0000 UTC,Reason:NoFrequentContainerdRestart,Message:containerd is functioning properly,},NodeCondition{Type:KernelDeadlock,Status:False,LastHeartbeatTime:2022-11-26 05:10:29 +0000 UTC,LastTransitionTime:2022-11-26 05:05:27 +0000 UTC,Reason:KernelHasNoDeadlock,Message:kernel has no deadlock,},NodeCondition{Type:ReadonlyFilesystem,Status:False,LastHeartbeatTime:2022-11-26 05:10:29 +0000 UTC,LastTransitionTime:2022-11-26 05:05:27 +0000 UTC,Reason:FilesystemIsNotReadOnly,Message:Filesystem is not read-only,},NodeCondition{Type:CorruptDockerOverlay2,Status:False,LastHeartbeatTime:2022-11-26 05:10:29 +0000 UTC,LastTransitionTime:2022-11-26 05:05:27 +0000 UTC,Reason:NoCorruptDockerOverlay2,Message:docker overlay2 is functioning properly,},NodeCondition{Type:FrequentUnregisterNetDevice,Status:False,LastHeartbeatTime:2022-11-26 05:10:29 +0000 UTC,LastTransitionTime:2022-11-26 05:05:27 +0000 UTC,Reason:NoFrequentUnregisterNetDevice,Message:node is functioning properly,},NodeCondition{Type:NetworkUnavailable,Status:False,LastHeartbeatTime:2022-11-26 05:05:39 +0000 UTC,LastTransitionTime:2022-11-26 05:05:39 +0000 UTC,Reason:RouteCreated,Message:RouteController created a route,},NodeCondition{Type:MemoryPressure,Status:False,LastHeartbeatTime:2022-11-26 05:10:46 +0000 UTC,LastTransitionTime:2022-11-26 05:05:22 +0000 UTC,Reason:KubeletHasSufficientMemory,Message:kubelet has sufficient memory available,},NodeCondition{Type:DiskPressure,Status:False,LastHeartbeatTime:2022-11-26 05:10:46 +0000 UTC,LastTransitionTime:2022-11-26 05:05:22 +0000 UTC,Reason:KubeletHasNoDiskPressure,Message:kubelet has no disk pressure,},NodeCondition{Type:PIDPressure,Status:False,LastHeartbeatTime:2022-11-26 05:10:46 +0000 UTC,LastTransitionTime:2022-11-26 05:05:22 +0000 UTC,Reason:KubeletHasSufficientPID,Message:kubelet has sufficient PID available,},NodeCondition{Type:Ready,Status:True,LastHeartbeatTime:2022-11-26 05:10:46 +0000 UTC,LastTransitionTime:2022-11-26 05:05:23 +0000 UTC,Reason:KubeletReady,Message:kubelet is posting ready status. AppArmor enabled,},},Addresses:[]NodeAddress{NodeAddress{Type:InternalIP,Address:10.138.0.5,},NodeAddress{Type:ExternalIP,Address:35.247.104.236,},NodeAddress{Type:InternalDNS,Address:bootstrap-e2e-minion-group-0975.c.k8s-jkns-e2e-gce-ci-reboot.internal,},NodeAddress{Type:Hostname,Address:bootstrap-e2e-minion-group-0975.c.k8s-jkns-e2e-gce-ci-reboot.internal,},},DaemonEndpoints:NodeDaemonEndpoints{KubeletEndpoint:DaemonEndpoint{Port:10250,},},NodeInfo:NodeSystemInfo{MachineID:b38385f7976a474936da82f5f565760a,SystemUUID:b38385f7-976a-4749-36da-82f5f565760a,BootID:0aa74273-9b8d-48ac-bad0-0a1805f010ea,KernelVersion:5.10.123+,OSImage:Container-Optimized OS from Google,ContainerRuntimeVersion:containerd://1.7.0-beta.0-149-gd06318622,KubeletVersion:v1.27.0-alpha.0.50+70617042976dc1,KubeProxyVersion:v1.27.0-alpha.0.50+70617042976dc1,OperatingSystem:linux,Architecture:amd64,},Images:[]ContainerImage{ContainerImage{Names:[registry.k8s.io/e2e-test-images/jessie-dnsutils@sha256:24aaf2626d6b27864c29de2097e8bbb840b3a414271bf7c8995e431e47d8408e registry.k8s.io/e2e-test-images/jessie-dnsutils:1.7],SizeBytes:112030336,},ContainerImage{Names:[registry.k8s.io/sig-storage/nfs-provisioner@sha256:e943bb77c7df05ebdc8c7888b2db289b13bf9f012d6a3a5a74f14d4d5743d439 registry.k8s.io/sig-storage/nfs-provisioner:v3.0.1],SizeBytes:90632047,},ContainerImage{Names:[registry.k8s.io/kube-proxy-amd64:v1.27.0-alpha.0.50_70617042976dc1],SizeBytes:67201736,},ContainerImage{Names:[registry.k8s.io/e2e-test-images/agnhost@sha256:16bbf38c463a4223d8cfe4da12bc61010b082a79b4bb003e2d3ba3ece5dd5f9e registry.k8s.io/e2e-test-images/agnhost:2.43],SizeBytes:51706353,},ContainerImage{Names:[gke.gcr.io/prometheus-to-sd@sha256:e739643c3939ba0b161425f45a1989eedfc4a3b166db9a7100863296b4c70510 gke.gcr.io/prometheus-to-sd:v0.11.1-gke.1],SizeBytes:48742566,},ContainerImage{Names:[registry.k8s.io/e2e-test-images/httpd@sha256:148b022f5c5da426fc2f3c14b5c0867e58ef05961510c84749ac1fddcb0fef22 registry.k8s.io/e2e-test-images/httpd:2.4.38-4],SizeBytes:40764257,},ContainerImage{Names:[registry.k8s.io/metrics-server/metrics-server@sha256:6385aec64bb97040a5e692947107b81e178555c7a5b71caa90d733e4130efc10 registry.k8s.io/metrics-server/metrics-server:v0.5.2],SizeBytes:26023008,},ContainerImage{Names:[registry.k8s.io/sig-storage/csi-provisioner@sha256:ee3b525d5b89db99da3b8eb521d9cd90cb6e9ef0fbb651e98bb37be78d36b5b8 registry.k8s.io/sig-storage/csi-provisioner:v3.3.0],SizeBytes:25491225,},ContainerImage{Names:[registry.k8s.io/sig-storage/csi-resizer@sha256:425d8f1b769398127767b06ed97ce62578a3179bcb99809ce93a1649e025ffe7 registry.k8s.io/sig-storage/csi-resizer:v1.6.0],SizeBytes:24148884,},ContainerImage{Names:[registry.k8s.io/sig-storage/csi-snapshotter@sha256:291334908ddf71a4661fd7f6d9d97274de8a5378a2b6fdfeb2ce73414a34f82f registry.k8s.io/sig-storage/csi-snapshotter:v6.1.0],SizeBytes:23881995,},ContainerImage{Names:[registry.k8s.io/sig-storage/csi-attacher@sha256:9a685020911e2725ad019dbce6e4a5ab93d51e3d4557f115e64343345e05781b registry.k8s.io/sig-storage/csi-attacher:v4.0.0],SizeBytes:23847201,},ContainerImage{Names:[registry.k8s.io/sig-storage/hostpathplugin@sha256:92257881c1d6493cf18299a24af42330f891166560047902b8d431fb66b01af5 registry.k8s.io/sig-storage/hostpathplugin:v1.9.0],SizeBytes:18758628,},ContainerImage{Names:[registry.k8s.io/autoscaling/addon-resizer@sha256:43f129b81d28f0fdd54de6d8e7eacd5728030782e03db16087fc241ad747d3d6 registry.k8s.io/autoscaling/addon-resizer:1.8.14],SizeBytes:10153852,},ContainerImage{Names:[registry.k8s.io/sig-storage/csi-node-driver-registrar@sha256:0103eee7c35e3e0b5cd8cdca9850dc71c793cdeb6669d8be7a89440da2d06ae4 registry.k8s.io/sig-storage/csi-node-driver-registrar:v2.5.1],SizeBytes:9133109,},ContainerImage{Names:[registry.k8s.io/sig-storage/livenessprobe@sha256:933940f13b3ea0abc62e656c1aa5c5b47c04b15d71250413a6b821bd0c58b94e registry.k8s.io/sig-storage/livenessprobe:v2.7.0],SizeBytes:8688564,},ContainerImage{Names:[registry.k8s.io/kas-network-proxy/proxy-agent@sha256:48f2a4ec3e10553a81b8dd1c6fa5fe4bcc9617f78e71c1ca89c6921335e2d7da registry.k8s.io/kas-network-proxy/proxy-agent:v0.0.33],SizeBytes:8512162,},ContainerImage{Names:[registry.k8s.io/metadata-proxy@sha256:e914645f22e946bce5165737e1b244e0a296ad1f0f81a9531adc57af2780978a registry.k8s.io/metadata-proxy:v0.1.12],SizeBytes:5301657,},ContainerImage{Names:[registry.k8s.io/e2e-test-images/busybox@sha256:2e0f836850e09b8b7cc937681d6194537a09fbd5f6b9e08f4d646a85128e8937 registry.k8s.io/e2e-test-images/busybox:1.29-4],SizeBytes:731990,},ContainerImage{Names:[registry.k8s.io/pause@sha256:9001185023633d17a2f98ff69b6ff2615b8ea02a825adffa40422f51dfdcde9d registry.k8s.io/pause:3.8],SizeBytes:311286,},},VolumesInUse:[],VolumesAttached:[]AttachedVolume{},Config:nil,},} Nov 26 05:11:57.833: INFO: Logging kubelet events for node bootstrap-e2e-minion-group-0975 Nov 26 05:11:57.888: INFO: Logging pods the kubelet thinks is on node bootstrap-e2e-minion-group-0975 Nov 26 05:11:57.979: INFO: pod-secrets-ae6e2110-abaa-4433-ab33-c417114f813f started at 2022-11-26 05:10:35 +0000 UTC (0+1 container statuses recorded) Nov 26 05:11:57.979: INFO: Container creates-volume-test ready: false, restart count 0 Nov 26 05:11:57.979: INFO: httpd started at 2022-11-26 05:09:49 +0000 UTC (0+1 container statuses recorded) Nov 26 05:11:57.979: INFO: Container httpd ready: true, restart count 0 Nov 26 05:11:57.979: INFO: netserver-0 started at 2022-11-26 05:11:06 +0000 UTC (0+1 container statuses recorded) Nov 26 05:11:57.979: INFO: Container webserver ready: true, restart count 2 Nov 26 05:11:57.979: INFO: pod-secrets-be23d693-e903-42f2-a4b6-dc83cf5645d2 started at 2022-11-26 05:07:34 +0000 UTC (0+1 container statuses recorded) Nov 26 05:11:57.979: INFO: Container creates-volume-test ready: false, restart count 0 Nov 26 05:11:57.979: INFO: test-container-pod started at 2022-11-26 05:11:10 +0000 UTC (0+1 container statuses recorded) Nov 26 05:11:57.979: INFO: Container webserver ready: true, restart count 0 Nov 26 05:11:57.979: INFO: external-provisioner-mqkmk started at 2022-11-26 05:11:34 +0000 UTC (0+1 container statuses recorded) Nov 26 05:11:57.979: INFO: Container nfs-provisioner ready: false, restart count 1 Nov 26 05:11:57.979: INFO: kube-proxy-bootstrap-e2e-minion-group-0975 started at 2022-11-26 05:05:23 +0000 UTC (0+1 container statuses recorded) Nov 26 05:11:57.979: INFO: Container kube-proxy ready: true, restart count 4 Nov 26 05:11:57.979: INFO: var-expansion-318f7926-d76e-4dae-8c51-56fecf6585b0 started at 2022-11-26 05:11:22 +0000 UTC (0+1 container statuses recorded) Nov 26 05:11:57.979: INFO: Container dapi-container ready: false, restart count 0 Nov 26 05:11:57.979: INFO: external-provisioner-ntc44 started at 2022-11-26 05:11:08 +0000 UTC (0+1 container statuses recorded) Nov 26 05:11:57.979: INFO: Container nfs-provisioner ready: true, restart count 0 Nov 26 05:11:57.979: INFO: pod-configmaps-5ce855f5-39dd-468a-b6b1-8fa57df85118 started at 2022-11-26 05:08:35 +0000 UTC (0+1 container statuses recorded) Nov 26 05:11:57.979: INFO: Container agnhost-container ready: false, restart count 0 Nov 26 05:11:57.979: INFO: netserver-0 started at 2022-11-26 05:10:47 +0000 UTC (0+1 container statuses recorded) Nov 26 05:11:57.979: INFO: Container webserver ready: false, restart count 3 Nov 26 05:11:57.979: INFO: netserver-0 started at 2022-11-26 05:11:29 +0000 UTC (0+1 container statuses recorded) Nov 26 05:11:57.979: INFO: Container webserver ready: true, restart count 0 Nov 26 05:11:57.979: INFO: netserver-0 started at 2022-11-26 05:09:51 +0000 UTC (0+1 container statuses recorded) Nov 26 05:11:57.979: INFO: Container webserver ready: true, restart count 3 Nov 26 05:11:57.979: INFO: forbid-27823991-z2w8j started at 2022-11-26 05:11:00 +0000 UTC (0+1 container statuses recorded) Nov 26 05:11:57.979: INFO: Container c ready: true, restart count 1 Nov 26 05:11:57.979: INFO: metadata-proxy-v0.1-mznrc started at 2022-11-26 05:05:24 +0000 UTC (0+2 container statuses recorded) Nov 26 05:11:57.979: INFO: Container metadata-proxy ready: true, restart count 0 Nov 26 05:11:57.979: INFO: Container prometheus-to-sd-exporter ready: true, restart count 0 Nov 26 05:11:57.979: INFO: konnectivity-agent-wfc29 started at 2022-11-26 05:05:39 +0000 UTC (0+1 container statuses recorded) Nov 26 05:11:57.979: INFO: Container konnectivity-agent ready: false, restart count 4 Nov 26 05:11:57.979: INFO: metrics-server-v0.5.2-867b8754b9-27xcg started at 2022-11-26 05:05:51 +0000 UTC (0+2 container statuses recorded) Nov 26 05:11:57.979: INFO: Container metrics-server ready: false, restart count 4 Nov 26 05:11:57.979: INFO: Container metrics-server-nanny ready: true, restart count 5 Nov 26 05:11:57.979: INFO: csi-hostpathplugin-0 started at 2022-11-26 05:09:53 +0000 UTC (0+7 container statuses recorded) Nov 26 05:11:57.979: INFO: Container csi-attacher ready: true, restart count 0 Nov 26 05:11:57.979: INFO: Container csi-provisioner ready: true, restart count 0 Nov 26 05:11:57.979: INFO: Container csi-resizer ready: true, restart count 0 Nov 26 05:11:57.979: INFO: Container csi-snapshotter ready: true, restart count 0 Nov 26 05:11:57.979: INFO: Container hostpath ready: true, restart count 0 Nov 26 05:11:57.979: INFO: Container liveness-probe ready: true, restart count 0 Nov 26 05:11:57.979: INFO: Container node-driver-registrar ready: true, restart count 0 Nov 26 05:11:57.979: INFO: affinity-lb-esipp-transition-8sg8w started at 2022-11-26 05:10:44 +0000 UTC (0+1 container statuses recorded) Nov 26 05:11:57.979: INFO: Container affinity-lb-esipp-transition ready: true, restart count 0 Nov 26 05:11:57.979: INFO: external-local-update-5xshb started at 2022-11-26 05:11:23 +0000 UTC (0+1 container statuses recorded) Nov 26 05:11:57.979: INFO: Container netexec ready: true, restart count 2 Nov 26 05:11:58.583: INFO: Latency metrics for node bootstrap-e2e-minion-group-0975 Nov 26 05:11:58.583: INFO: Logging node info for node bootstrap-e2e-minion-group-rwsn Nov 26 05:11:58.682: INFO: Node Info: &Node{ObjectMeta:{bootstrap-e2e-minion-group-rwsn adedf225-b462-4305-a60c-9042d31ca1d2 4961 0 2022-11-26 05:05:15 +0000 UTC <nil> <nil> map[beta.kubernetes.io/arch:amd64 beta.kubernetes.io/instance-type:n1-standard-2 beta.kubernetes.io/os:linux cloud.google.com/metadata-proxy-ready:true failure-domain.beta.kubernetes.io/region:us-west1 failure-domain.beta.kubernetes.io/zone:us-west1-b kubernetes.io/arch:amd64 kubernetes.io/hostname:bootstrap-e2e-minion-group-rwsn kubernetes.io/os:linux node.kubernetes.io/instance-type:n1-standard-2 topology.hostpath.csi/node:bootstrap-e2e-minion-group-rwsn topology.kubernetes.io/region:us-west1 topology.kubernetes.io/zone:us-west1-b] map[node.alpha.kubernetes.io/ttl:0 volumes.kubernetes.io/controller-managed-attach-detach:true] [] [] [{kubelet Update v1 2022-11-26 05:05:15 +0000 UTC FieldsV1 {"f:metadata":{"f:annotations":{".":{},"f:volumes.kubernetes.io/controller-managed-attach-detach":{}},"f:labels":{".":{},"f:beta.kubernetes.io/arch":{},"f:beta.kubernetes.io/instance-type":{},"f:beta.kubernetes.io/os":{},"f:cloud.google.com/metadata-proxy-ready":{},"f:failure-domain.beta.kubernetes.io/region":{},"f:failure-domain.beta.kubernetes.io/zone":{},"f:kubernetes.io/arch":{},"f:kubernetes.io/hostname":{},"f:kubernetes.io/os":{},"f:node.kubernetes.io/instance-type":{},"f:topology.kubernetes.io/region":{},"f:topology.kubernetes.io/zone":{}}},"f:spec":{"f:providerID":{}}} } {kube-controller-manager Update v1 2022-11-26 05:05:17 +0000 UTC FieldsV1 {"f:metadata":{"f:annotations":{"f:node.alpha.kubernetes.io/ttl":{}}},"f:spec":{"f:podCIDR":{},"f:podCIDRs":{".":{},"v:\"10.64.0.0/24\"":{}}}} } {kube-controller-manager Update v1 2022-11-26 05:08:09 +0000 UTC FieldsV1 {"f:status":{"f:conditions":{"k:{\"type\":\"NetworkUnavailable\"}":{"f:lastHeartbeatTime":{},"f:lastTransitionTime":{},"f:message":{},"f:reason":{},"f:status":{}}},"f:volumesAttached":{}}} status} {node-problem-detector Update v1 2022-11-26 05:10:19 +0000 UTC FieldsV1 {"f:status":{"f:conditions":{"k:{\"type\":\"CorruptDockerOverlay2\"}":{".":{},"f:lastHeartbeatTime":{},"f:lastTransitionTime":{},"f:message":{},"f:reason":{},"f:status":{},"f:type":{}},"k:{\"type\":\"FrequentContainerdRestart\"}":{".":{},"f:lastHeartbeatTime":{},"f:lastTransitionTime":{},"f:message":{},"f:reason":{},"f:status":{},"f:type":{}},"k:{\"type\":\"FrequentDockerRestart\"}":{".":{},"f:lastHeartbeatTime":{},"f:lastTransitionTime":{},"f:message":{},"f:reason":{},"f:status":{},"f:type":{}},"k:{\"type\":\"FrequentKubeletRestart\"}":{".":{},"f:lastHeartbeatTime":{},"f:lastTransitionTime":{},"f:message":{},"f:reason":{},"f:status":{},"f:type":{}},"k:{\"type\":\"FrequentUnregisterNetDevice\"}":{".":{},"f:lastHeartbeatTime":{},"f:lastTransitionTime":{},"f:message":{},"f:reason":{},"f:status":{},"f:type":{}},"k:{\"type\":\"KernelDeadlock\"}":{".":{},"f:lastHeartbeatTime":{},"f:lastTransitionTime":{},"f:message":{},"f:reason":{},"f:status":{},"f:type":{}},"k:{\"type\":\"ReadonlyFilesystem\"}":{".":{},"f:lastHeartbeatTime":{},"f:lastTransitionTime":{},"f:message":{},"f:reason":{},"f:status":{},"f:type":{}}}}} status} {kubelet Update v1 2022-11-26 05:11:17 +0000 UTC FieldsV1 {"f:metadata":{"f:labels":{"f:topology.hostpath.csi/node":{}}},"f:status":{"f:conditions":{"k:{\"type\":\"DiskPressure\"}":{"f:lastHeartbeatTime":{}},"k:{\"type\":\"MemoryPressure\"}":{"f:lastHeartbeatTime":{}},"k:{\"type\":\"PIDPressure\"}":{"f:lastHeartbeatTime":{}},"k:{\"type\":\"Ready\"}":{"f:lastHeartbeatTime":{},"f:lastTransitionTime":{},"f:message":{},"f:reason":{},"f:status":{}}},"f:images":{},"f:volumesInUse":{}}} status}]},Spec:NodeSpec{PodCIDR:10.64.0.0/24,DoNotUseExternalID:,ProviderID:gce://k8s-jkns-e2e-gce-ci-reboot/us-west1-b/bootstrap-e2e-minion-group-rwsn,Unschedulable:false,Taints:[]Taint{},ConfigSource:nil,PodCIDRs:[10.64.0.0/24],},Status:NodeStatus{Capacity:ResourceList{attachable-volumes-gce-pd: {{127 0} {<nil>} 127 DecimalSI},cpu: {{2 0} {<nil>} 2 DecimalSI},ephemeral-storage: {{101203873792 0} {<nil>} 98831908Ki BinarySI},hugepages-1Gi: {{0 0} {<nil>} 0 DecimalSI},hugepages-2Mi: {{0 0} {<nil>} 0 DecimalSI},memory: {{7815438336 0} {<nil>} BinarySI},pods: {{110 0} {<nil>} 110 DecimalSI},},Allocatable:ResourceList{attachable-volumes-gce-pd: {{127 0} {<nil>} 127 DecimalSI},cpu: {{2 0} {<nil>} 2 DecimalSI},ephemeral-storage: {{91083486262 0} {<nil>} 91083486262 DecimalSI},hugepages-1Gi: {{0 0} {<nil>} 0 DecimalSI},hugepages-2Mi: {{0 0} {<nil>} 0 DecimalSI},memory: {{7553294336 0} {<nil>} BinarySI},pods: {{110 0} {<nil>} 110 DecimalSI},},Phase:,Conditions:[]NodeCondition{NodeCondition{Type:ReadonlyFilesystem,Status:False,LastHeartbeatTime:2022-11-26 05:10:19 +0000 UTC,LastTransitionTime:2022-11-26 05:05:18 +0000 UTC,Reason:FilesystemIsNotReadOnly,Message:Filesystem is not read-only,},NodeCondition{Type:CorruptDockerOverlay2,Status:False,LastHeartbeatTime:2022-11-26 05:10:19 +0000 UTC,LastTransitionTime:2022-11-26 05:05:18 +0000 UTC,Reason:NoCorruptDockerOverlay2,Message:docker overlay2 is functioning properly,},NodeCondition{Type:FrequentUnregisterNetDevice,Status:False,LastHeartbeatTime:2022-11-26 05:10:19 +0000 UTC,LastTransitionTime:2022-11-26 05:05:18 +0000 UTC,Reason:NoFrequentUnregisterNetDevice,Message:node is functioning properly,},NodeCondition{Type:FrequentKubeletRestart,Status:False,LastHeartbeatTime:2022-11-26 05:10:19 +0000 UTC,LastTransitionTime:2022-11-26 05:05:18 +0000 UTC,Reason:NoFrequentKubeletRestart,Message:kubelet is functioning properly,},NodeCondition{Type:FrequentDockerRestart,Status:False,LastHeartbeatTime:2022-11-26 05:10:19 +0000 UTC,LastTransitionTime:2022-11-26 05:05:18 +0000 UTC,Reason:NoFrequentDockerRestart,Message:docker is functioning properly,},NodeCondition{Type:FrequentContainerdRestart,Status:False,LastHeartbeatTime:2022-11-26 05:10:19 +0000 UTC,LastTransitionTime:2022-11-26 05:05:18 +0000 UTC,Reason:NoFrequentContainerdRestart,Message:containerd is functioning properly,},NodeCondition{Type:KernelDeadlock,Status:False,LastHeartbeatTime:2022-11-26 05:10:19 +0000 UTC,LastTransitionTime:2022-11-26 05:05:18 +0000 UTC,Reason:KernelHasNoDeadlock,Message:kernel has no deadlock,},NodeCondition{Type:NetworkUnavailable,Status:False,LastHeartbeatTime:2022-11-26 05:05:25 +0000 UTC,LastTransitionTime:2022-11-26 05:05:25 +0000 UTC,Reason:RouteCreated,Message:RouteController created a route,},NodeCondition{Type:MemoryPressure,Status:False,LastHeartbeatTime:2022-11-26 05:08:19 +0000 UTC,LastTransitionTime:2022-11-26 05:05:15 +0000 UTC,Reason:KubeletHasSufficientMemory,Message:kubelet has sufficient memory available,},NodeCondition{Type:DiskPressure,Status:False,LastHeartbeatTime:2022-11-26 05:08:19 +0000 UTC,LastTransitionTime:2022-11-26 05:05:15 +0000 UTC,Reason:KubeletHasNoDiskPressure,Message:kubelet has no disk pressure,},NodeCondition{Type:PIDPressure,Status:False,LastHeartbeatTime:2022-11-26 05:08:19 +0000 UTC,LastTransitionTime:2022-11-26 05:05:15 +0000 UTC,Reason:KubeletHasSufficientPID,Message:kubelet has sufficient PID available,},NodeCondition{Type:Ready,Status:True,LastHeartbeatTime:2022-11-26 05:08:19 +0000 UTC,LastTransitionTime:2022-11-26 05:05:17 +0000 UTC,Reason:KubeletReady,Message:kubelet is posting ready status. AppArmor enabled,},},Addresses:[]NodeAddress{NodeAddress{Type:InternalIP,Address:10.138.0.4,},NodeAddress{Type:ExternalIP,Address:35.230.27.45,},NodeAddress{Type:InternalDNS,Address:bootstrap-e2e-minion-group-rwsn.c.k8s-jkns-e2e-gce-ci-reboot.internal,},NodeAddress{Type:Hostname,Address:bootstrap-e2e-minion-group-rwsn.c.k8s-jkns-e2e-gce-ci-reboot.internal,},},DaemonEndpoints:NodeDaemonEndpoints{KubeletEndpoint:DaemonEndpoint{Port:10250,},},NodeInfo:NodeSystemInfo{MachineID:69e22ce1281686d6cde1b3211b34b8af,SystemUUID:69e22ce1-2816-86d6-cde1-b3211b34b8af,BootID:79e9e1fd-2a39-42ec-b72b-72836d213862,KernelVersion:5.10.123+,OSImage:Container-Optimized OS from Google,ContainerRuntimeVersion:containerd://1.7.0-beta.0-149-gd06318622,KubeletVersion:v1.27.0-alpha.0.50+70617042976dc1,KubeProxyVersion:v1.27.0-alpha.0.50+70617042976dc1,OperatingSystem:linux,Architecture:amd64,},Images:[]ContainerImage{ContainerImage{Names:[registry.k8s.io/sig-storage/nfs-provisioner@sha256:e943bb77c7df05ebdc8c7888b2db289b13bf9f012d6a3a5a74f14d4d5743d439 registry.k8s.io/sig-storage/nfs-provisioner:v3.0.1],SizeBytes:90632047,},ContainerImage{Names:[registry.k8s.io/kube-proxy-amd64:v1.27.0-alpha.0.50_70617042976dc1],SizeBytes:67201736,},ContainerImage{Names:[registry.k8s.io/e2e-test-images/agnhost@sha256:16bbf38c463a4223d8cfe4da12bc61010b082a79b4bb003e2d3ba3ece5dd5f9e registry.k8s.io/e2e-test-images/agnhost:2.43],SizeBytes:51706353,},ContainerImage{Names:[gke.gcr.io/prometheus-to-sd@sha256:e739643c3939ba0b161425f45a1989eedfc4a3b166db9a7100863296b4c70510 gke.gcr.io/prometheus-to-sd:v0.11.1-gke.1],SizeBytes:48742566,},ContainerImage{Names:[registry.k8s.io/metrics-server/metrics-server@sha256:6385aec64bb97040a5e692947107b81e178555c7a5b71caa90d733e4130efc10 registry.k8s.io/metrics-server/metrics-server:v0.5.2],SizeBytes:26023008,},ContainerImage{Names:[registry.k8s.io/sig-storage/csi-provisioner@sha256:ee3b525d5b89db99da3b8eb521d9cd90cb6e9ef0fbb651e98bb37be78d36b5b8 registry.k8s.io/sig-storage/csi-provisioner:v3.3.0],SizeBytes:25491225,},ContainerImage{Names:[registry.k8s.io/sig-storage/csi-resizer@sha256:425d8f1b769398127767b06ed97ce62578a3179bcb99809ce93a1649e025ffe7 registry.k8s.io/sig-storage/csi-resizer:v1.6.0],SizeBytes:24148884,},ContainerImage{Names:[registry.k8s.io/sig-storage/csi-snapshotter@sha256:291334908ddf71a4661fd7f6d9d97274de8a5378a2b6fdfeb2ce73414a34f82f registry.k8s.io/sig-storage/csi-snapshotter:v6.1.0],SizeBytes:23881995,},ContainerImage{Names:[registry.k8s.io/sig-storage/csi-attacher@sha256:9a685020911e2725ad019dbce6e4a5ab93d51e3d4557f115e64343345e05781b registry.k8s.io/sig-storage/csi-attacher:v4.0.0],SizeBytes:23847201,},ContainerImage{Names:[registry.k8s.io/sig-storage/snapshot-controller@sha256:823c75d0c45d1427f6d850070956d9ca657140a7bbf828381541d1d808475280 registry.k8s.io/sig-storage/snapshot-controller:v6.1.0],SizeBytes:22620891,},ContainerImage{Names:[registry.k8s.io/sig-storage/hostpathplugin@sha256:92257881c1d6493cf18299a24af42330f891166560047902b8d431fb66b01af5 registry.k8s.io/sig-storage/hostpathplugin:v1.9.0],SizeBytes:18758628,},ContainerImage{Names:[registry.k8s.io/cpa/cluster-proportional-autoscaler@sha256:fd636b33485c7826fb20ef0688a83ee0910317dbb6c0c6f3ad14661c1db25def registry.k8s.io/cpa/cluster-proportional-autoscaler:1.8.4],SizeBytes:15209393,},ContainerImage{Names:[registry.k8s.io/coredns/coredns@sha256:8e352a029d304ca7431c6507b56800636c321cb52289686a581ab70aaa8a2e2a registry.k8s.io/coredns/coredns:v1.9.3],SizeBytes:14837849,},ContainerImage{Names:[registry.k8s.io/autoscaling/addon-resizer@sha256:43f129b81d28f0fdd54de6d8e7eacd5728030782e03db16087fc241ad747d3d6 registry.k8s.io/autoscaling/addon-resizer:1.8.14],SizeBytes:10153852,},ContainerImage{Names:[registry.k8s.io/sig-storage/csi-node-driver-registrar@sha256:0103eee7c35e3e0b5cd8cdca9850dc71c793cdeb6669d8be7a89440da2d06ae4 registry.k8s.io/sig-storage/csi-node-driver-registrar:v2.5.1],SizeBytes:9133109,},ContainerImage{Names:[registry.k8s.io/sig-storage/livenessprobe@sha256:933940f13b3ea0abc62e656c1aa5c5b47c04b15d71250413a6b821bd0c58b94e registry.k8s.io/sig-storage/livenessprobe:v2.7.0],SizeBytes:8688564,},ContainerImage{Names:[registry.k8s.io/kas-network-proxy/proxy-agent@sha256:48f2a4ec3e10553a81b8dd1c6fa5fe4bcc9617f78e71c1ca89c6921335e2d7da registry.k8s.io/kas-network-proxy/proxy-agent:v0.0.33],SizeBytes:8512162,},ContainerImage{Names:[registry.k8s.io/networking/ingress-gce-404-server-with-metrics-amd64@sha256:7eb7b3cee4d33c10c49893ad3c386232b86d4067de5251294d4c620d6e072b93 registry.k8s.io/networking/ingress-gce-404-server-with-metrics-amd64:v1.10.11],SizeBytes:6463068,},ContainerImage{Names:[registry.k8s.io/metadata-proxy@sha256:e914645f22e946bce5165737e1b244e0a296ad1f0f81a9531adc57af2780978a registry.k8s.io/metadata-proxy:v0.1.12],SizeBytes:5301657,},ContainerImage{Names:[registry.k8s.io/e2e-test-images/busybox@sha256:c318242786b139d18676b1c09a0ad7f15fc17f8f16a5b2e625cd0dc8c9703daf registry.k8s.io/e2e-test-images/busybox:1.29-2],SizeBytes:732424,},ContainerImage{Names:[registry.k8s.io/e2e-test-images/busybox@sha256:2e0f836850e09b8b7cc937681d6194537a09fbd5f6b9e08f4d646a85128e8937 registry.k8s.io/e2e-test-images/busybox:1.29-4],SizeBytes:731990,},ContainerImage{Names:[registry.k8s.io/pause@sha256:9001185023633d17a2f98ff69b6ff2615b8ea02a825adffa40422f51dfdcde9d registry.k8s.io/pause:3.8],SizeBytes:311286,},},VolumesInUse:[kubernetes.io/csi/csi-hostpath-provisioning-6984^3b0457df-6d48-11ed-9e4d-d626cdd9a50a],VolumesAttached:[]AttachedVolume{AttachedVolume{Name:kubernetes.io/csi/csi-hostpath-provisioning-6984^3b0457df-6d48-11ed-9e4d-d626cdd9a50a,DevicePath:,},},Config:nil,},} Nov 26 05:11:58.682: INFO: Logging kubelet events for node bootstrap-e2e-minion-group-rwsn Nov 26 05:11:58.741: INFO: Logging pods the kubelet thinks is on node bootstrap-e2e-minion-group-rwsn Nov 26 05:11:58.841: INFO: pod-subpath-test-dynamicpv-h6m7 started at 2022-11-26 05:07:31 +0000 UTC (1+1 container statuses recorded) Nov 26 05:11:58.841: INFO: Init container init-volume-dynamicpv-h6m7 ready: true, restart count 0 Nov 26 05:11:58.841: INFO: Container test-container-subpath-dynamicpv-h6m7 ready: false, restart count 0 Nov 26 05:11:58.841: INFO: coredns-6d97d5ddb-xk6v4 started at 2022-11-26 05:05:25 +0000 UTC (0+1 container statuses recorded) Nov 26 05:11:58.841: INFO: Container coredns ready: false, restart count 4 Nov 26 05:11:58.841: INFO: konnectivity-agent-sxg5q started at 2022-11-26 05:05:25 +0000 UTC (0+1 container statuses recorded) Nov 26 05:11:58.841: INFO: Container konnectivity-agent ready: true, restart count 4 Nov 26 05:11:58.841: INFO: csi-hostpathplugin-0 started at 2022-11-26 05:07:18 +0000 UTC (0+7 container statuses recorded) Nov 26 05:11:58.841: INFO: Container csi-attacher ready: false, restart count 3 Nov 26 05:11:58.841: INFO: Container csi-provisioner ready: false, restart count 3 Nov 26 05:11:58.841: INFO: Container csi-resizer ready: false, restart count 3 Nov 26 05:11:58.841: INFO: Container csi-snapshotter ready: false, restart count 3 Nov 26 05:11:58.841: INFO: Container hostpath ready: false, restart count 3 Nov 26 05:11:58.841: INFO: Container liveness-probe ready: false, restart count 3 Nov 26 05:11:58.841: INFO: Container node-driver-registrar ready: false, restart count 3 Nov 26 05:11:58.841: INFO: kube-dns-autoscaler-5f6455f985-xktw8 started at 2022-11-26 05:05:25 +0000 UTC (0+1 container statuses recorded) Nov 26 05:11:58.841: INFO: Container autoscaler ready: false, restart count 4 Nov 26 05:11:58.841: INFO: coredns-6d97d5ddb-p2xpt started at 2022-11-26 05:05:31 +0000 UTC (0+1 container statuses recorded) Nov 26 05:11:58.841: INFO: Container coredns ready: false, restart count 5 Nov 26 05:11:58.841: INFO: affinity-lb-esipp-transition-6h4g4 started at 2022-11-26 05:10:44 +0000 UTC (0+1 container statuses recorded) Nov 26 05:11:58.841: INFO: Container affinity-lb-esipp-transition ready: true, restart count 2 Nov 26 05:11:58.841: INFO: csi-mockplugin-0 started at 2022-11-26 05:08:11 +0000 UTC (0+4 container statuses recorded) Nov 26 05:11:58.841: INFO: Container busybox ready: true, restart count 3 Nov 26 05:11:58.841: INFO: Container csi-provisioner ready: true, restart count 3 Nov 26 05:11:58.841: INFO: Container driver-registrar ready: false, restart count 3 Nov 26 05:11:58.841: INFO: Container mock ready: false, restart count 3 Nov 26 05:11:58.841: INFO: netserver-1 started at 2022-11-26 05:11:06 +0000 UTC (0+1 container statuses recorded) Nov 26 05:11:58.841: INFO: Container webserver ready: false, restart count 2 Nov 26 05:11:58.841: INFO: kube-proxy-bootstrap-e2e-minion-group-rwsn started at 2022-11-26 05:05:15 +0000 UTC (0+1 container statuses recorded) Nov 26 05:11:58.841: INFO: Container kube-proxy ready: true, restart count 3 Nov 26 05:11:58.841: INFO: l7-default-backend-8549d69d99-8cfzv started at 2022-11-26 05:05:25 +0000 UTC (0+1 container statuses recorded) Nov 26 05:11:58.841: INFO: Container default-http-backend ready: true, restart count 0 Nov 26 05:11:58.841: INFO: volume-snapshot-controller-0 started at 2022-11-26 05:05:25 +0000 UTC (0+1 container statuses recorded) Nov 26 05:11:58.841: INFO: Container volume-snapshot-controller ready: true, restart count 3 Nov 26 05:11:58.841: INFO: csi-hostpathplugin-0 started at 2022-11-26 05:07:19 +0000 UTC (0+7 container statuses recorded) Nov 26 05:11:58.841: INFO: Container csi-attacher ready: false, restart count 5 Nov 26 05:11:58.841: INFO: Container csi-provisioner ready: false, restart count 5 Nov 26 05:11:58.841: INFO: Container csi-resizer ready: false, restart count 5 Nov 26 05:11:58.841: INFO: Container csi-snapshotter ready: false, restart count 5 Nov 26 05:11:58.841: INFO: Container hostpath ready: false, restart count 5 Nov 26 05:11:58.841: INFO: Container liveness-probe ready: false, restart count 5 Nov 26 05:11:58.841: INFO: Container node-driver-registrar ready: false, restart count 5 Nov 26 05:11:58.841: INFO: netserver-1 started at 2022-11-26 05:11:29 +0000 UTC (0+1 container statuses recorded) Nov 26 05:11:58.841: INFO: Container webserver ready: false, restart count 2 Nov 26 05:11:58.841: INFO: metadata-proxy-v0.1-qlckk started at 2022-11-26 05:05:16 +0000 UTC (0+2 container statuses recorded) Nov 26 05:11:58.841: INFO: Container metadata-proxy ready: true, restart count 0 Nov 26 05:11:58.841: INFO: Container prometheus-to-sd-exporter ready: true, restart count 0 Nov 26 05:11:58.841: INFO: netserver-1 started at 2022-11-26 05:09:51 +0000 UTC (0+1 container statuses recorded) Nov 26 05:11:58.841: INFO: Container webserver ready: true, restart count 4 Nov 26 05:11:58.841: INFO: csi-hostpathplugin-0 started at 2022-11-26 05:10:11 +0000 UTC (0+7 container statuses recorded) Nov 26 05:11:58.841: INFO: Container csi-attacher ready: false, restart count 2 Nov 26 05:11:58.841: INFO: Container csi-provisioner ready: false, restart count 2 Nov 26 05:11:58.841: INFO: Container csi-resizer ready: false, restart count 2 Nov 26 05:11:58.841: INFO: Container csi-snapshotter ready: false, restart count 2 Nov 26 05:11:58.841: INFO: Container hostpath ready: false, restart count 2 Nov 26 05:11:58.841: INFO: Container liveness-probe ready: false, restart count 2 Nov 26 05:11:58.841: INFO: Container node-driver-registrar ready: false, restart count 2 Nov 26 05:11:58.841: INFO: netserver-1 started at 2022-11-26 05:10:47 +0000 UTC (0+1 container statuses recorded) Nov 26 05:11:58.841: INFO: Container webserver ready: true, restart count 1 Nov 26 05:11:59.155: INFO: Latency metrics for node bootstrap-e2e-minion-group-rwsn Nov 26 05:11:59.155: INFO: Logging node info for node bootstrap-e2e-minion-group-vhdj Nov 26 05:11:59.208: INFO: Node Info: &Node{ObjectMeta:{bootstrap-e2e-minion-group-vhdj 0d27fd16-9ec8-4428-bb1e-97fead6c7a66 4538 0 2022-11-26 05:05:23 +0000 UTC <nil> <nil> map[beta.kubernetes.io/arch:amd64 beta.kubernetes.io/instance-type:n1-standard-2 beta.kubernetes.io/os:linux cloud.google.com/metadata-proxy-ready:true failure-domain.beta.kubernetes.io/region:us-west1 failure-domain.beta.kubernetes.io/zone:us-west1-b kubernetes.io/arch:amd64 kubernetes.io/hostname:bootstrap-e2e-minion-group-vhdj kubernetes.io/os:linux node.kubernetes.io/instance-type:n1-standard-2 topology.hostpath.csi/node:bootstrap-e2e-minion-group-vhdj topology.kubernetes.io/region:us-west1 topology.kubernetes.io/zone:us-west1-b] map[csi.volume.kubernetes.io/nodeid:{"csi-hostpath-multivolume-9910":"bootstrap-e2e-minion-group-vhdj","csi-hostpath-provisioning-7617":"bootstrap-e2e-minion-group-vhdj"} node.alpha.kubernetes.io/ttl:0 volumes.kubernetes.io/controller-managed-attach-detach:true] [] [] [{kubelet Update v1 2022-11-26 05:05:23 +0000 UTC FieldsV1 {"f:metadata":{"f:annotations":{".":{},"f:volumes.kubernetes.io/controller-managed-attach-detach":{}},"f:labels":{".":{},"f:beta.kubernetes.io/arch":{},"f:beta.kubernetes.io/instance-type":{},"f:beta.kubernetes.io/os":{},"f:cloud.google.com/metadata-proxy-ready":{},"f:failure-domain.beta.kubernetes.io/region":{},"f:failure-domain.beta.kubernetes.io/zone":{},"f:kubernetes.io/arch":{},"f:kubernetes.io/hostname":{},"f:kubernetes.io/os":{},"f:node.kubernetes.io/instance-type":{},"f:topology.kubernetes.io/region":{},"f:topology.kubernetes.io/zone":{}}},"f:spec":{"f:providerID":{}}} } {kube-controller-manager Update v1 2022-11-26 05:05:24 +0000 UTC FieldsV1 {"f:metadata":{"f:annotations":{"f:node.alpha.kubernetes.io/ttl":{}}},"f:spec":{"f:podCIDR":{},"f:podCIDRs":{".":{},"v:\"10.64.3.0/24\"":{}}}} } {node-problem-detector Update v1 2022-11-26 05:10:27 +0000 UTC FieldsV1 {"f:status":{"f:conditions":{"k:{\"type\":\"CorruptDockerOverlay2\"}":{".":{},"f:lastHeartbeatTime":{},"f:lastTransitionTime":{},"f:message":{},"f:reason":{},"f:status":{},"f:type":{}},"k:{\"type\":\"FrequentContainerdRestart\"}":{".":{},"f:lastHeartbeatTime":{},"f:lastTransitionTime":{},"f:message":{},"f:reason":{},"f:status":{},"f:type":{}},"k:{\"type\":\"FrequentDockerRestart\"}":{".":{},"f:lastHeartbeatTime":{},"f:lastTransitionTime":{},"f:message":{},"f:reason":{},"f:status":{},"f:type":{}},"k:{\"type\":\"FrequentKubeletRestart\"}":{".":{},"f:lastHeartbeatTime":{},"f:lastTransitionTime":{},"f:message":{},"f:reason":{},"f:status":{},"f:type":{}},"k:{\"type\":\"FrequentUnregisterNetDevice\"}":{".":{},"f:lastHeartbeatTime":{},"f:lastTransitionTime":{},"f:message":{},"f:reason":{},"f:status":{},"f:type":{}},"k:{\"type\":\"KernelDeadlock\"}":{".":{},"f:lastHeartbeatTime":{},"f:lastTransitionTime":{},"f:message":{},"f:reason":{},"f:status":{},"f:type":{}},"k:{\"type\":\"ReadonlyFilesystem\"}":{".":{},"f:lastHeartbeatTime":{},"f:lastTransitionTime":{},"f:message":{},"f:reason":{},"f:status":{},"f:type":{}}}}} status} {kube-controller-manager Update v1 2022-11-26 05:10:55 +0000 UTC FieldsV1 {"f:status":{"f:conditions":{"k:{\"type\":\"NetworkUnavailable\"}":{"f:lastHeartbeatTime":{},"f:lastTransitionTime":{},"f:message":{},"f:reason":{},"f:status":{}}}}} status} {kubelet Update v1 2022-11-26 05:11:07 +0000 UTC FieldsV1 {"f:metadata":{"f:annotations":{"f:csi.volume.kubernetes.io/nodeid":{}},"f:labels":{"f:topology.hostpath.csi/node":{}}},"f:status":{"f:conditions":{"k:{\"type\":\"DiskPressure\"}":{"f:lastHeartbeatTime":{}},"k:{\"type\":\"MemoryPressure\"}":{"f:lastHeartbeatTime":{}},"k:{\"type\":\"PIDPressure\"}":{"f:lastHeartbeatTime":{}},"k:{\"type\":\"Ready\"}":{"f:lastHeartbeatTime":{},"f:lastTransitionTime":{},"f:message":{},"f:reason":{},"f:status":{}}},"f:images":{}}} status}]},Spec:NodeSpec{PodCIDR:10.64.3.0/24,DoNotUseExternalID:,ProviderID:gce://k8s-jkns-e2e-gce-ci-reboot/us-west1-b/bootstrap-e2e-minion-group-vhdj,Unschedulable:false,Taints:[]Taint{},ConfigSource:nil,PodCIDRs:[10.64.3.0/24],},Status:NodeStatus{Capacity:ResourceList{attachable-volumes-gce-pd: {{127 0} {<nil>} 127 DecimalSI},cpu: {{2 0} {<nil>} 2 DecimalSI},ephemeral-storage: {{101203873792 0} {<nil>} 98831908Ki BinarySI},hugepages-1Gi: {{0 0} {<nil>} 0 DecimalSI},hugepages-2Mi: {{0 0} {<nil>} 0 DecimalSI},memory: {{7815438336 0} {<nil>} BinarySI},pods: {{110 0} {<nil>} 110 DecimalSI},},Allocatable:ResourceList{attachable-volumes-gce-pd: {{127 0} {<nil>} 127 DecimalSI},cpu: {{2 0} {<nil>} 2 DecimalSI},ephemeral-storage: {{91083486262 0} {<nil>} 91083486262 DecimalSI},hugepages-1Gi: {{0 0} {<nil>} 0 DecimalSI},hugepages-2Mi: {{0 0} {<nil>} 0 DecimalSI},memory: {{7553294336 0} {<nil>} BinarySI},pods: {{110 0} {<nil>} 110 DecimalSI},},Phase:,Conditions:[]NodeCondition{NodeCondition{Type:CorruptDockerOverlay2,Status:False,LastHeartbeatTime:2022-11-26 05:10:27 +0000 UTC,LastTransitionTime:2022-11-26 05:05:26 +0000 UTC,Reason:NoCorruptDockerOverlay2,Message:docker overlay2 is functioning properly,},NodeCondition{Type:FrequentUnregisterNetDevice,Status:False,LastHeartbeatTime:2022-11-26 05:10:27 +0000 UTC,LastTransitionTime:2022-11-26 05:05:26 +0000 UTC,Reason:NoFrequentUnregisterNetDevice,Message:node is functioning properly,},NodeCondition{Type:FrequentKubeletRestart,Status:False,LastHeartbeatTime:2022-11-26 05:10:27 +0000 UTC,LastTransitionTime:2022-11-26 05:05:26 +0000 UTC,Reason:NoFrequentKubeletRestart,Message:kubelet is functioning properly,},NodeCondition{Type:FrequentDockerRestart,Status:False,LastHeartbeatTime:2022-11-26 05:10:27 +0000 UTC,LastTransitionTime:2022-11-26 05:05:26 +0000 UTC,Reason:NoFrequentDockerRestart,Message:docker is functioning properly,},NodeCondition{Type:FrequentContainerdRestart,Status:False,LastHeartbeatTime:2022-11-26 05:10:27 +0000 UTC,LastTransitionTime:2022-11-26 05:05:26 +0000 UTC,Reason:NoFrequentContainerdRestart,Message:containerd is functioning properly,},NodeCondition{Type:KernelDeadlock,Status:False,LastHeartbeatTime:2022-11-26 05:10:27 +0000 UTC,LastTransitionTime:2022-11-26 05:05:26 +0000 UTC,Reason:KernelHasNoDeadlock,Message:kernel has no deadlock,},NodeCondition{Type:ReadonlyFilesystem,Status:False,LastHeartbeatTime:2022-11-26 05:10:27 +0000 UTC,LastTransitionTime:2022-11-26 05:05:26 +0000 UTC,Reason:FilesystemIsNotReadOnly,Message:Filesystem is not read-only,},NodeCondition{Type:NetworkUnavailable,Status:False,LastHeartbeatTime:2022-11-26 05:05:39 +0000 UTC,LastTransitionTime:2022-11-26 05:05:39 +0000 UTC,Reason:RouteCreated,Message:RouteController created a route,},NodeCondition{Type:MemoryPressure,Status:False,LastHeartbeatTime:2022-11-26 05:11:07 +0000 UTC,LastTransitionTime:2022-11-26 05:05:23 +0000 UTC,Reason:KubeletHasSufficientMemory,Message:kubelet has sufficient memory available,},NodeCondition{Type:DiskPressure,Status:False,LastHeartbeatTime:2022-11-26 05:11:07 +0000 UTC,LastTransitionTime:2022-11-26 05:05:23 +0000 UTC,Reason:KubeletHasNoDiskPressure,Message:kubelet has no disk pressure,},NodeCondition{Type:PIDPressure,Status:False,LastHeartbeatTime:2022-11-26 05:11:07 +0000 UTC,LastTransitionTime:2022-11-26 05:05:23 +0000 UTC,Reason:KubeletHasSufficientPID,Message:kubelet has sufficient PID available,},NodeCondition{Type:Ready,Status:True,LastHeartbeatTime:2022-11-26 05:11:07 +0000 UTC,LastTransitionTime:2022-11-26 05:05:24 +0000 UTC,Reason:KubeletReady,Message:kubelet is posting ready status. AppArmor enabled,},},Addresses:[]NodeAddress{NodeAddress{Type:InternalIP,Address:10.138.0.3,},NodeAddress{Type:ExternalIP,Address:34.83.94.215,},NodeAddress{Type:InternalDNS,Address:bootstrap-e2e-minion-group-vhdj.c.k8s-jkns-e2e-gce-ci-reboot.internal,},NodeAddress{Type:Hostname,Address:bootstrap-e2e-minion-group-vhdj.c.k8s-jkns-e2e-gce-ci-reboot.internal,},},DaemonEndpoints:NodeDaemonEndpoints{KubeletEndpoint:DaemonEndpoint{Port:10250,},},NodeInfo:NodeSystemInfo{MachineID:31b0953ad3e740aeb7fc1a89edb5435a,SystemUUID:31b0953a-d3e7-40ae-b7fc-1a89edb5435a,BootID:14c054d0-bda9-4ddf-b9f6-e1742be33d25,KernelVersion:5.10.123+,OSImage:Container-Optimized OS from Google,ContainerRuntimeVersion:containerd://1.7.0-beta.0-149-gd06318622,KubeletVersion:v1.27.0-alpha.0.50+70617042976dc1,KubeProxyVersion:v1.27.0-alpha.0.50+70617042976dc1,OperatingSystem:linux,Architecture:amd64,},Images:[]ContainerImage{ContainerImage{Names:[registry.k8s.io/e2e-test-images/jessie-dnsutils@sha256:24aaf2626d6b27864c29de2097e8bbb840b3a414271bf7c8995e431e47d8408e registry.k8s.io/e2e-test-images/jessie-dnsutils:1.7],SizeBytes:112030336,},ContainerImage{Names:[registry.k8s.io/sig-storage/nfs-provisioner@sha256:e943bb77c7df05ebdc8c7888b2db289b13bf9f012d6a3a5a74f14d4d5743d439 registry.k8s.io/sig-storage/nfs-provisioner:v3.0.1],SizeBytes:90632047,},ContainerImage{Names:[registry.k8s.io/kube-proxy-amd64:v1.27.0-alpha.0.50_70617042976dc1],SizeBytes:67201736,},ContainerImage{Names:[registry.k8s.io/e2e-test-images/agnhost@sha256:16bbf38c463a4223d8cfe4da12bc61010b082a79b4bb003e2d3ba3ece5dd5f9e registry.k8s.io/e2e-test-images/agnhost:2.43],SizeBytes:51706353,},ContainerImage{Names:[gke.gcr.io/prometheus-to-sd@sha256:e739643c3939ba0b161425f45a1989eedfc4a3b166db9a7100863296b4c70510 gke.gcr.io/prometheus-to-sd:v0.11.1-gke.1],SizeBytes:48742566,},ContainerImage{Names:[registry.k8s.io/sig-storage/csi-provisioner@sha256:ee3b525d5b89db99da3b8eb521d9cd90cb6e9ef0fbb651e98bb37be78d36b5b8 registry.k8s.io/sig-storage/csi-provisioner:v3.3.0],SizeBytes:25491225,},ContainerImage{Names:[registry.k8s.io/sig-storage/csi-resizer@sha256:425d8f1b769398127767b06ed97ce62578a3179bcb99809ce93a1649e025ffe7 registry.k8s.io/sig-storage/csi-resizer:v1.6.0],SizeBytes:24148884,},ContainerImage{Names:[registry.k8s.io/sig-storage/csi-snapshotter@sha256:291334908ddf71a4661fd7f6d9d97274de8a5378a2b6fdfeb2ce73414a34f82f registry.k8s.io/sig-storage/csi-snapshotter:v6.1.0],SizeBytes:23881995,},ContainerImage{Names:[registry.k8s.io/sig-storage/csi-attacher@sha256:9a685020911e2725ad019dbce6e4a5ab93d51e3d4557f115e64343345e05781b registry.k8s.io/sig-storage/csi-attacher:v4.0.0],SizeBytes:23847201,},ContainerImage{Names:[registry.k8s.io/sig-storage/hostpathplugin@sha256:92257881c1d6493cf18299a24af42330f891166560047902b8d431fb66b01af5 registry.k8s.io/sig-storage/hostpathplugin:v1.9.0],SizeBytes:18758628,},ContainerImage{Names:[registry.k8s.io/sig-storage/csi-node-driver-registrar@sha256:0103eee7c35e3e0b5cd8cdca9850dc71c793cdeb6669d8be7a89440da2d06ae4 registry.k8s.io/sig-storage/csi-node-driver-registrar:v2.5.1],SizeBytes:9133109,},ContainerImage{Names:[registry.k8s.io/sig-storage/livenessprobe@sha256:933940f13b3ea0abc62e656c1aa5c5b47c04b15d71250413a6b821bd0c58b94e registry.k8s.io/sig-storage/livenessprobe:v2.7.0],SizeBytes:8688564,},ContainerImage{Names:[registry.k8s.io/kas-network-proxy/proxy-agent@sha256:48f2a4ec3e10553a81b8dd1c6fa5fe4bcc9617f78e71c1ca89c6921335e2d7da registry.k8s.io/kas-network-proxy/proxy-agent:v0.0.33],SizeBytes:8512162,},ContainerImage{Names:[registry.k8s.io/metadata-proxy@sha256:e914645f22e946bce5165737e1b244e0a296ad1f0f81a9531adc57af2780978a registry.k8s.io/metadata-proxy:v0.1.12],SizeBytes:5301657,},ContainerImage{Names:[registry.k8s.io/e2e-test-images/busybox@sha256:2e0f836850e09b8b7cc937681d6194537a09fbd5f6b9e08f4d646a85128e8937 registry.k8s.io/e2e-test-images/busybox:1.29-4],SizeBytes:731990,},ContainerImage{Names:[registry.k8s.io/pause@sha256:9001185023633d17a2f98ff69b6ff2615b8ea02a825adffa40422f51dfdcde9d registry.k8s.io/pause:3.8],SizeBytes:311286,},},VolumesInUse:[],VolumesAttached:[]AttachedVolume{},Config:nil,},} Nov 26 05:11:59.209: INFO: Logging kubelet events for node bootstrap-e2e-minion-group-vhdj Nov 26 05:11:59.277: INFO: Logging pods the kubelet thinks is on node bootstrap-e2e-minion-group-vhdj Nov 26 05:11:59.372: INFO: csi-hostpathplugin-0 started at 2022-11-26 05:08:31 +0000 UTC (0+7 container statuses recorded) Nov 26 05:11:59.372: INFO: Container csi-attacher ready: true, restart count 1 Nov 26 05:11:59.372: INFO: Container csi-provisioner ready: true, restart count 1 Nov 26 05:11:59.372: INFO: Container csi-resizer ready: true, restart count 1 Nov 26 05:11:59.372: INFO: Container csi-snapshotter ready: true, restart count 1 Nov 26 05:11:59.372: INFO: Container hostpath ready: true, restart count 1 Nov 26 05:11:59.372: INFO: Container liveness-probe ready: true, restart count 1 Nov 26 05:11:59.372: INFO: Container node-driver-registrar ready: true, restart count 1 Nov 26 05:11:59.372: INFO: netserver-2 started at 2022-11-26 05:11:06 +0000 UTC (0+1 container statuses recorded) Nov 26 05:11:59.372: INFO: Container webserver ready: true, restart count 0 Nov 26 05:11:59.372: INFO: netserver-2 started at 2022-11-26 05:11:29 +0000 UTC (0+1 container statuses recorded) Nov 26 05:11:59.372: INFO: Container webserver ready: true, restart count 1 Nov 26 05:11:59.372: INFO: kube-proxy-bootstrap-e2e-minion-group-vhdj started at 2022-11-26 05:05:23 +0000 UTC (0+1 container statuses recorded) Nov 26 05:11:59.372: INFO: Container kube-proxy ready: false, restart count 4 Nov 26 05:11:59.372: INFO: pod-9fe7e984-9a6a-4bb0-9c65-0628a6a2f792 started at 2022-11-26 05:08:13 +0000 UTC (0+1 container statuses recorded) Nov 26 05:11:59.372: INFO: Container write-pod ready: false, restart count 0 Nov 26 05:11:59.372: INFO: failure-3 started at 2022-11-26 05:10:01 +0000 UTC (0+1 container statuses recorded) Nov 26 05:11:59.372: INFO: Container failure-3 ready: true, restart count 2 Nov 26 05:11:59.372: INFO: netserver-2 started at 2022-11-26 05:10:47 +0000 UTC (0+1 container statuses recorded) Nov 26 05:11:59.372: INFO: Container webserver ready: true, restart count 1 Nov 26 05:11:59.372: INFO: metadata-proxy-v0.1-wn5l5 started at 2022-11-26 05:05:24 +0000 UTC (0+2 container statuses recorded) Nov 26 05:11:59.372: INFO: Container metadata-proxy ready: true, restart count 0 Nov 26 05:11:59.372: INFO: Container prometheus-to-sd-exporter ready: true, restart count 0 Nov 26 05:11:59.372: INFO: pod-subpath-test-dynamicpv-4nfp started at 2022-11-26 05:11:17 +0000 UTC (1+2 container statuses recorded) Nov 26 05:11:59.372: INFO: Init container init-volume-dynamicpv-4nfp ready: true, restart count 0 Nov 26 05:11:59.372: INFO: Container test-container-subpath-dynamicpv-4nfp ready: true, restart count 2 Nov 26 05:11:59.372: INFO: Container test-container-volume-dynamicpv-4nfp ready: true, restart count 2 Nov 26 05:11:59.372: INFO: konnectivity-agent-nt4fd started at 2022-11-26 05:05:39 +0000 UTC (0+1 container statuses recorded) Nov 26 05:11:59.372: INFO: Container konnectivity-agent ready: true, restart count 4 Nov 26 05:11:59.372: INFO: affinity-lb-esipp-transition-ml7rt started at 2022-11-26 05:10:44 +0000 UTC (0+1 container statuses recorded) Nov 26 05:11:59.372: INFO: Container affinity-lb-esipp-transition ready: true, restart count 1 Nov 26 05:11:59.372: INFO: netserver-2 started at 2022-11-26 05:09:51 +0000 UTC (0+1 container statuses recorded) Nov 26 05:11:59.372: INFO: Container webserver ready: true, restart count 2 Nov 26 05:11:59.372: INFO: csi-hostpathplugin-0 started at 2022-11-26 05:10:50 +0000 UTC (0+7 container statuses recorded) Nov 26 05:11:59.372: INFO: Container csi-attacher ready: true, restart count 0 Nov 26 05:11:59.372: INFO: Container csi-provisioner ready: true, restart count 0 Nov 26 05:11:59.372: INFO: Container csi-resizer ready: true, restart count 0 Nov 26 05:11:59.372: INFO: Container csi-snapshotter ready: true, restart count 0 Nov 26 05:11:59.372: INFO: Container hostpath ready: true, restart count 0 Nov 26 05:11:59.372: INFO: Container liveness-probe ready: true, restart count 0 Nov 26 05:11:59.372: INFO: Container node-driver-registrar ready: true, restart count 0 Nov 26 05:11:59.372: INFO: back-off-cap started at 2022-11-26 05:07:40 +0000 UTC (0+1 container statuses recorded) Nov 26 05:11:59.372: INFO: Container back-off-cap ready: false, restart count 5 Nov 26 05:11:59.372: INFO: mutability-test-5jd94 started at 2022-11-26 05:07:54 +0000 UTC (0+1 container statuses recorded) Nov 26 05:11:59.372: INFO: Container netexec ready: true, restart count 2 Nov 26 05:11:59.672: INFO: Latency metrics for node bootstrap-e2e-minion-group-vhdj [DeferCleanup (Each)] [sig-network] LoadBalancers ESIPP [Slow] tear down framework | framework.go:193 STEP: Destroying namespace "esipp-2129" for this suite. 11/26/22 05:11:59.672
Filter through log files | View test history on testgrid
go run hack/e2e.go -v --test --test_args='--ginkgo.focus=Kubernetes\se2e\ssuite\s\[It\]\s\[sig\-network\]\sLoadBalancers\sESIPP\s\[Slow\]\sshould\swork\sfor\stype\=LoadBalancer$'
test/e2e/network/loadbalancer.go:1272 k8s.io/kubernetes/test/e2e/network.glob..func20.3() test/e2e/network/loadbalancer.go:1272 +0xd8from junit_01.xml
[BeforeEach] [sig-network] LoadBalancers ESIPP [Slow] set up framework | framework.go:178 STEP: Creating a kubernetes client 11/26/22 05:17:32.371 Nov 26 05:17:32.371: INFO: >>> kubeConfig: /workspace/.kube/config STEP: Building a namespace api object, basename esipp 11/26/22 05:17:32.373 STEP: Waiting for a default service account to be provisioned in namespace 11/26/22 05:17:32.552 STEP: Waiting for kube-root-ca.crt to be provisioned in namespace 11/26/22 05:17:32.65 [BeforeEach] [sig-network] LoadBalancers ESIPP [Slow] test/e2e/framework/metrics/init/init.go:31 [BeforeEach] [sig-network] LoadBalancers ESIPP [Slow] test/e2e/network/loadbalancer.go:1250 [It] should work for type=LoadBalancer test/e2e/network/loadbalancer.go:1266 STEP: creating a service esipp-2671/external-local-lb with type=LoadBalancer 11/26/22 05:17:33.023 STEP: setting ExternalTrafficPolicy=Local 11/26/22 05:17:33.023 STEP: waiting for loadbalancer for service esipp-2671/external-local-lb 11/26/22 05:17:33.34 Nov 26 05:17:33.340: INFO: Waiting up to 15m0s for service "external-local-lb" to have a LoadBalancer ------------------------------ Progress Report for Ginkgo Process #11 Automatically polling progress: [sig-network] LoadBalancers ESIPP [Slow] should work for type=LoadBalancer (Spec Runtime: 5m0.652s) test/e2e/network/loadbalancer.go:1266 In [It] (Node Runtime: 5m0s) test/e2e/network/loadbalancer.go:1266 At [By Step] waiting for loadbalancer for service esipp-2671/external-local-lb (Step Runtime: 4m59.683s) test/e2e/framework/service/jig.go:260 Spec Goroutine goroutine 2087 [select] k8s.io/kubernetes/vendor/k8s.io/apimachinery/pkg/util/wait.WaitForWithContext({0x7fe0bc8, 0xc00013a000}, 0xc001dd11a0, 0x2fdb16a?) vendor/k8s.io/apimachinery/pkg/util/wait/wait.go:660 k8s.io/kubernetes/vendor/k8s.io/apimachinery/pkg/util/wait.poll({0x7fe0bc8, 0xc00013a000}, 0x28?, 0x2fd9d05?, 0x20?) vendor/k8s.io/apimachinery/pkg/util/wait/wait.go:596 k8s.io/kubernetes/vendor/k8s.io/apimachinery/pkg/util/wait.PollImmediateWithContext({0x7fe0bc8, 0xc00013a000}, 0xc001534f00?, 0xc0055adb78?, 0x262a967?) vendor/k8s.io/apimachinery/pkg/util/wait/wait.go:528 k8s.io/kubernetes/vendor/k8s.io/apimachinery/pkg/util/wait.PollImmediate(0xc0010da770?, 0x7fa7740?, 0xc000214bc0?) vendor/k8s.io/apimachinery/pkg/util/wait/wait.go:514 > k8s.io/kubernetes/test/e2e/framework/service.(*TestJig).waitForCondition(0xc0047adef0, 0x4?, {0x7600fe2, 0x14}, 0x7895b68) test/e2e/framework/service/jig.go:631 > k8s.io/kubernetes/test/e2e/framework/service.(*TestJig).WaitForLoadBalancer(0xc0047adef0, 0x41?) test/e2e/framework/service/jig.go:582 > k8s.io/kubernetes/test/e2e/framework/service.(*TestJig).CreateLoadBalancerService(0xc0047adef0, 0x6aba880?, 0xc0055ade28) test/e2e/framework/service/jig.go:261 > k8s.io/kubernetes/test/e2e/framework/service.(*TestJig).CreateOnlyLocalLoadBalancerService(0xc0047adef0, 0xc002aabd40?, 0x1, 0xa?) test/e2e/framework/service/jig.go:222 > k8s.io/kubernetes/test/e2e/network.glob..func20.3() test/e2e/network/loadbalancer.go:1271 k8s.io/kubernetes/vendor/github.com/onsi/ginkgo/v2/internal.extractBodyFunction.func3({0x2d591ce, 0xc0004cfe00}) vendor/github.com/onsi/ginkgo/v2/internal/node.go:449 k8s.io/kubernetes/vendor/github.com/onsi/ginkgo/v2/internal.(*Suite).runNode.func2() vendor/github.com/onsi/ginkgo/v2/internal/suite.go:750 k8s.io/kubernetes/vendor/github.com/onsi/ginkgo/v2/internal.(*Suite).runNode vendor/github.com/onsi/ginkgo/v2/internal/suite.go:738 ------------------------------ ------------------------------ Progress Report for Ginkgo Process #11 Automatically polling progress: [sig-network] LoadBalancers ESIPP [Slow] should work for type=LoadBalancer (Spec Runtime: 5m20.655s) test/e2e/network/loadbalancer.go:1266 In [It] (Node Runtime: 5m20.003s) test/e2e/network/loadbalancer.go:1266 At [By Step] waiting for loadbalancer for service esipp-2671/external-local-lb (Step Runtime: 5m19.686s) test/e2e/framework/service/jig.go:260 Spec Goroutine goroutine 2087 [select] k8s.io/kubernetes/vendor/k8s.io/apimachinery/pkg/util/wait.WaitForWithContext({0x7fe0bc8, 0xc00013a000}, 0xc001dd11a0, 0x2fdb16a?) vendor/k8s.io/apimachinery/pkg/util/wait/wait.go:660 k8s.io/kubernetes/vendor/k8s.io/apimachinery/pkg/util/wait.poll({0x7fe0bc8, 0xc00013a000}, 0x28?, 0x2fd9d05?, 0x20?) vendor/k8s.io/apimachinery/pkg/util/wait/wait.go:596 k8s.io/kubernetes/vendor/k8s.io/apimachinery/pkg/util/wait.PollImmediateWithContext({0x7fe0bc8, 0xc00013a000}, 0xc001534f00?, 0xc0055adb78?, 0x262a967?) vendor/k8s.io/apimachinery/pkg/util/wait/wait.go:528 k8s.io/kubernetes/vendor/k8s.io/apimachinery/pkg/util/wait.PollImmediate(0xc0010da770?, 0x7fa7740?, 0xc000214bc0?) vendor/k8s.io/apimachinery/pkg/util/wait/wait.go:514 > k8s.io/kubernetes/test/e2e/framework/service.(*TestJig).waitForCondition(0xc0047adef0, 0x4?, {0x7600fe2, 0x14}, 0x7895b68) test/e2e/framework/service/jig.go:631 > k8s.io/kubernetes/test/e2e/framework/service.(*TestJig).WaitForLoadBalancer(0xc0047adef0, 0x41?) test/e2e/framework/service/jig.go:582 > k8s.io/kubernetes/test/e2e/framework/service.(*TestJig).CreateLoadBalancerService(0xc0047adef0, 0x6aba880?, 0xc0055ade28) test/e2e/framework/service/jig.go:261 > k8s.io/kubernetes/test/e2e/framework/service.(*TestJig).CreateOnlyLocalLoadBalancerService(0xc0047adef0, 0xc002aabd40?, 0x1, 0xa?) test/e2e/framework/service/jig.go:222 > k8s.io/kubernetes/test/e2e/network.glob..func20.3() test/e2e/network/loadbalancer.go:1271 k8s.io/kubernetes/vendor/github.com/onsi/ginkgo/v2/internal.extractBodyFunction.func3({0x2d591ce, 0xc0004cfe00}) vendor/github.com/onsi/ginkgo/v2/internal/node.go:449 k8s.io/kubernetes/vendor/github.com/onsi/ginkgo/v2/internal.(*Suite).runNode.func2() vendor/github.com/onsi/ginkgo/v2/internal/suite.go:750 k8s.io/kubernetes/vendor/github.com/onsi/ginkgo/v2/internal.(*Suite).runNode vendor/github.com/onsi/ginkgo/v2/internal/suite.go:738 ------------------------------ ------------------------------ Progress Report for Ginkgo Process #11 Automatically polling progress: [sig-network] LoadBalancers ESIPP [Slow] should work for type=LoadBalancer (Spec Runtime: 5m40.657s) test/e2e/network/loadbalancer.go:1266 In [It] (Node Runtime: 5m40.006s) test/e2e/network/loadbalancer.go:1266 At [By Step] waiting for loadbalancer for service esipp-2671/external-local-lb (Step Runtime: 5m39.688s) test/e2e/framework/service/jig.go:260 Spec Goroutine goroutine 2087 [select] k8s.io/kubernetes/vendor/k8s.io/apimachinery/pkg/util/wait.WaitForWithContext({0x7fe0bc8, 0xc00013a000}, 0xc001dd11a0, 0x2fdb16a?) vendor/k8s.io/apimachinery/pkg/util/wait/wait.go:660 k8s.io/kubernetes/vendor/k8s.io/apimachinery/pkg/util/wait.poll({0x7fe0bc8, 0xc00013a000}, 0x28?, 0x2fd9d05?, 0x20?) vendor/k8s.io/apimachinery/pkg/util/wait/wait.go:596 k8s.io/kubernetes/vendor/k8s.io/apimachinery/pkg/util/wait.PollImmediateWithContext({0x7fe0bc8, 0xc00013a000}, 0xc001534f00?, 0xc0055adb78?, 0x262a967?) vendor/k8s.io/apimachinery/pkg/util/wait/wait.go:528 k8s.io/kubernetes/vendor/k8s.io/apimachinery/pkg/util/wait.PollImmediate(0xc0010da770?, 0x7fa7740?, 0xc000214bc0?) vendor/k8s.io/apimachinery/pkg/util/wait/wait.go:514 > k8s.io/kubernetes/test/e2e/framework/service.(*TestJig).waitForCondition(0xc0047adef0, 0x4?, {0x7600fe2, 0x14}, 0x7895b68) test/e2e/framework/service/jig.go:631 > k8s.io/kubernetes/test/e2e/framework/service.(*TestJig).WaitForLoadBalancer(0xc0047adef0, 0x41?) test/e2e/framework/service/jig.go:582 > k8s.io/kubernetes/test/e2e/framework/service.(*TestJig).CreateLoadBalancerService(0xc0047adef0, 0x6aba880?, 0xc0055ade28) test/e2e/framework/service/jig.go:261 > k8s.io/kubernetes/test/e2e/framework/service.(*TestJig).CreateOnlyLocalLoadBalancerService(0xc0047adef0, 0xc002aabd40?, 0x1, 0xa?) test/e2e/framework/service/jig.go:222 > k8s.io/kubernetes/test/e2e/network.glob..func20.3() test/e2e/network/loadbalancer.go:1271 k8s.io/kubernetes/vendor/github.com/onsi/ginkgo/v2/internal.extractBodyFunction.func3({0x2d591ce, 0xc0004cfe00}) vendor/github.com/onsi/ginkgo/v2/internal/node.go:449 k8s.io/kubernetes/vendor/github.com/onsi/ginkgo/v2/internal.(*Suite).runNode.func2() vendor/github.com/onsi/ginkgo/v2/internal/suite.go:750 k8s.io/kubernetes/vendor/github.com/onsi/ginkgo/v2/internal.(*Suite).runNode vendor/github.com/onsi/ginkgo/v2/internal/suite.go:738 ------------------------------ ------------------------------ Progress Report for Ginkgo Process #11 Automatically polling progress: [sig-network] LoadBalancers ESIPP [Slow] should work for type=LoadBalancer (Spec Runtime: 6m0.66s) test/e2e/network/loadbalancer.go:1266 In [It] (Node Runtime: 6m0.008s) test/e2e/network/loadbalancer.go:1266 At [By Step] waiting for loadbalancer for service esipp-2671/external-local-lb (Step Runtime: 5m59.69s) test/e2e/framework/service/jig.go:260 Spec Goroutine goroutine 2087 [select] k8s.io/kubernetes/vendor/k8s.io/apimachinery/pkg/util/wait.WaitForWithContext({0x7fe0bc8, 0xc00013a000}, 0xc001dd11a0, 0x2fdb16a?) vendor/k8s.io/apimachinery/pkg/util/wait/wait.go:660 k8s.io/kubernetes/vendor/k8s.io/apimachinery/pkg/util/wait.poll({0x7fe0bc8, 0xc00013a000}, 0x28?, 0x2fd9d05?, 0x20?) vendor/k8s.io/apimachinery/pkg/util/wait/wait.go:596 k8s.io/kubernetes/vendor/k8s.io/apimachinery/pkg/util/wait.PollImmediateWithContext({0x7fe0bc8, 0xc00013a000}, 0xc001534f00?, 0xc0055adb78?, 0x262a967?) vendor/k8s.io/apimachinery/pkg/util/wait/wait.go:528 k8s.io/kubernetes/vendor/k8s.io/apimachinery/pkg/util/wait.PollImmediate(0xc0010da770?, 0x7fa7740?, 0xc000214bc0?) vendor/k8s.io/apimachinery/pkg/util/wait/wait.go:514 > k8s.io/kubernetes/test/e2e/framework/service.(*TestJig).waitForCondition(0xc0047adef0, 0x4?, {0x7600fe2, 0x14}, 0x7895b68) test/e2e/framework/service/jig.go:631 > k8s.io/kubernetes/test/e2e/framework/service.(*TestJig).WaitForLoadBalancer(0xc0047adef0, 0x41?) test/e2e/framework/service/jig.go:582 > k8s.io/kubernetes/test/e2e/framework/service.(*TestJig).CreateLoadBalancerService(0xc0047adef0, 0x6aba880?, 0xc0055ade28) test/e2e/framework/service/jig.go:261 > k8s.io/kubernetes/test/e2e/framework/service.(*TestJig).CreateOnlyLocalLoadBalancerService(0xc0047adef0, 0xc002aabd40?, 0x1, 0xa?) test/e2e/framework/service/jig.go:222 > k8s.io/kubernetes/test/e2e/network.glob..func20.3() test/e2e/network/loadbalancer.go:1271 k8s.io/kubernetes/vendor/github.com/onsi/ginkgo/v2/internal.extractBodyFunction.func3({0x2d591ce, 0xc0004cfe00}) vendor/github.com/onsi/ginkgo/v2/internal/node.go:449 k8s.io/kubernetes/vendor/github.com/onsi/ginkgo/v2/internal.(*Suite).runNode.func2() vendor/github.com/onsi/ginkgo/v2/internal/suite.go:750 k8s.io/kubernetes/vendor/github.com/onsi/ginkgo/v2/internal.(*Suite).runNode vendor/github.com/onsi/ginkgo/v2/internal/suite.go:738 ------------------------------ ------------------------------ Progress Report for Ginkgo Process #11 Automatically polling progress: [sig-network] LoadBalancers ESIPP [Slow] should work for type=LoadBalancer (Spec Runtime: 6m20.662s) test/e2e/network/loadbalancer.go:1266 In [It] (Node Runtime: 6m20.01s) test/e2e/network/loadbalancer.go:1266 At [By Step] waiting for loadbalancer for service esipp-2671/external-local-lb (Step Runtime: 6m19.693s) test/e2e/framework/service/jig.go:260 Spec Goroutine goroutine 2087 [select] k8s.io/kubernetes/vendor/k8s.io/apimachinery/pkg/util/wait.WaitForWithContext({0x7fe0bc8, 0xc00013a000}, 0xc001dd11a0, 0x2fdb16a?) vendor/k8s.io/apimachinery/pkg/util/wait/wait.go:660 k8s.io/kubernetes/vendor/k8s.io/apimachinery/pkg/util/wait.poll({0x7fe0bc8, 0xc00013a000}, 0x28?, 0x2fd9d05?, 0x20?) vendor/k8s.io/apimachinery/pkg/util/wait/wait.go:596 k8s.io/kubernetes/vendor/k8s.io/apimachinery/pkg/util/wait.PollImmediateWithContext({0x7fe0bc8, 0xc00013a000}, 0xc001534f00?, 0xc0055adb78?, 0x262a967?) vendor/k8s.io/apimachinery/pkg/util/wait/wait.go:528 k8s.io/kubernetes/vendor/k8s.io/apimachinery/pkg/util/wait.PollImmediate(0xc0010da770?, 0x7fa7740?, 0xc000214bc0?) vendor/k8s.io/apimachinery/pkg/util/wait/wait.go:514 > k8s.io/kubernetes/test/e2e/framework/service.(*TestJig).waitForCondition(0xc0047adef0, 0x4?, {0x7600fe2, 0x14}, 0x7895b68) test/e2e/framework/service/jig.go:631 > k8s.io/kubernetes/test/e2e/framework/service.(*TestJig).WaitForLoadBalancer(0xc0047adef0, 0x41?) test/e2e/framework/service/jig.go:582 > k8s.io/kubernetes/test/e2e/framework/service.(*TestJig).CreateLoadBalancerService(0xc0047adef0, 0x6aba880?, 0xc0055ade28) test/e2e/framework/service/jig.go:261 > k8s.io/kubernetes/test/e2e/framework/service.(*TestJig).CreateOnlyLocalLoadBalancerService(0xc0047adef0, 0xc002aabd40?, 0x1, 0xa?) test/e2e/framework/service/jig.go:222 > k8s.io/kubernetes/test/e2e/network.glob..func20.3() test/e2e/network/loadbalancer.go:1271 k8s.io/kubernetes/vendor/github.com/onsi/ginkgo/v2/internal.extractBodyFunction.func3({0x2d591ce, 0xc0004cfe00}) vendor/github.com/onsi/ginkgo/v2/internal/node.go:449 k8s.io/kubernetes/vendor/github.com/onsi/ginkgo/v2/internal.(*Suite).runNode.func2() vendor/github.com/onsi/ginkgo/v2/internal/suite.go:750 k8s.io/kubernetes/vendor/github.com/onsi/ginkgo/v2/internal.(*Suite).runNode vendor/github.com/onsi/ginkgo/v2/internal/suite.go:738 ------------------------------ ------------------------------ Progress Report for Ginkgo Process #11 Automatically polling progress: [sig-network] LoadBalancers ESIPP [Slow] should work for type=LoadBalancer (Spec Runtime: 6m40.667s) test/e2e/network/loadbalancer.go:1266 In [It] (Node Runtime: 6m40.015s) test/e2e/network/loadbalancer.go:1266 At [By Step] waiting for loadbalancer for service esipp-2671/external-local-lb (Step Runtime: 6m39.698s) test/e2e/framework/service/jig.go:260 Spec Goroutine goroutine 2087 [select] k8s.io/kubernetes/vendor/k8s.io/apimachinery/pkg/util/wait.WaitForWithContext({0x7fe0bc8, 0xc00013a000}, 0xc001dd11a0, 0x2fdb16a?) vendor/k8s.io/apimachinery/pkg/util/wait/wait.go:660 k8s.io/kubernetes/vendor/k8s.io/apimachinery/pkg/util/wait.poll({0x7fe0bc8, 0xc00013a000}, 0x28?, 0x2fd9d05?, 0x20?) vendor/k8s.io/apimachinery/pkg/util/wait/wait.go:596 k8s.io/kubernetes/vendor/k8s.io/apimachinery/pkg/util/wait.PollImmediateWithContext({0x7fe0bc8, 0xc00013a000}, 0xc001534f00?, 0xc0055adb78?, 0x262a967?) vendor/k8s.io/apimachinery/pkg/util/wait/wait.go:528 k8s.io/kubernetes/vendor/k8s.io/apimachinery/pkg/util/wait.PollImmediate(0xc0010da770?, 0x7fa7740?, 0xc000214bc0?) vendor/k8s.io/apimachinery/pkg/util/wait/wait.go:514 > k8s.io/kubernetes/test/e2e/framework/service.(*TestJig).waitForCondition(0xc0047adef0, 0x4?, {0x7600fe2, 0x14}, 0x7895b68) test/e2e/framework/service/jig.go:631 > k8s.io/kubernetes/test/e2e/framework/service.(*TestJig).WaitForLoadBalancer(0xc0047adef0, 0x41?) test/e2e/framework/service/jig.go:582 > k8s.io/kubernetes/test/e2e/framework/service.(*TestJig).CreateLoadBalancerService(0xc0047adef0, 0x6aba880?, 0xc0055ade28) test/e2e/framework/service/jig.go:261 > k8s.io/kubernetes/test/e2e/framework/service.(*TestJig).CreateOnlyLocalLoadBalancerService(0xc0047adef0, 0xc002aabd40?, 0x1, 0xa?) test/e2e/framework/service/jig.go:222 > k8s.io/kubernetes/test/e2e/network.glob..func20.3() test/e2e/network/loadbalancer.go:1271 k8s.io/kubernetes/vendor/github.com/onsi/ginkgo/v2/internal.extractBodyFunction.func3({0x2d591ce, 0xc0004cfe00}) vendor/github.com/onsi/ginkgo/v2/internal/node.go:449 k8s.io/kubernetes/vendor/github.com/onsi/ginkgo/v2/internal.(*Suite).runNode.func2() vendor/github.com/onsi/ginkgo/v2/internal/suite.go:750 k8s.io/kubernetes/vendor/github.com/onsi/ginkgo/v2/internal.(*Suite).runNode vendor/github.com/onsi/ginkgo/v2/internal/suite.go:738 ------------------------------ ------------------------------ Progress Report for Ginkgo Process #11 Automatically polling progress: [sig-network] LoadBalancers ESIPP [Slow] should work for type=LoadBalancer (Spec Runtime: 7m0.67s) test/e2e/network/loadbalancer.go:1266 In [It] (Node Runtime: 7m0.018s) test/e2e/network/loadbalancer.go:1266 At [By Step] waiting for loadbalancer for service esipp-2671/external-local-lb (Step Runtime: 6m59.7s) test/e2e/framework/service/jig.go:260 Spec Goroutine goroutine 2087 [select] k8s.io/kubernetes/vendor/k8s.io/apimachinery/pkg/util/wait.WaitForWithContext({0x7fe0bc8, 0xc00013a000}, 0xc001dd11a0, 0x2fdb16a?) vendor/k8s.io/apimachinery/pkg/util/wait/wait.go:660 k8s.io/kubernetes/vendor/k8s.io/apimachinery/pkg/util/wait.poll({0x7fe0bc8, 0xc00013a000}, 0x28?, 0x2fd9d05?, 0x20?) vendor/k8s.io/apimachinery/pkg/util/wait/wait.go:596 k8s.io/kubernetes/vendor/k8s.io/apimachinery/pkg/util/wait.PollImmediateWithContext({0x7fe0bc8, 0xc00013a000}, 0xc001534f00?, 0xc0055adb78?, 0x262a967?) vendor/k8s.io/apimachinery/pkg/util/wait/wait.go:528 k8s.io/kubernetes/vendor/k8s.io/apimachinery/pkg/util/wait.PollImmediate(0xc0010da770?, 0x7fa7740?, 0xc000214bc0?) vendor/k8s.io/apimachinery/pkg/util/wait/wait.go:514 > k8s.io/kubernetes/test/e2e/framework/service.(*TestJig).waitForCondition(0xc0047adef0, 0x4?, {0x7600fe2, 0x14}, 0x7895b68) test/e2e/framework/service/jig.go:631 > k8s.io/kubernetes/test/e2e/framework/service.(*TestJig).WaitForLoadBalancer(0xc0047adef0, 0x41?) test/e2e/framework/service/jig.go:582 > k8s.io/kubernetes/test/e2e/framework/service.(*TestJig).CreateLoadBalancerService(0xc0047adef0, 0x6aba880?, 0xc0055ade28) test/e2e/framework/service/jig.go:261 > k8s.io/kubernetes/test/e2e/framework/service.(*TestJig).CreateOnlyLocalLoadBalancerService(0xc0047adef0, 0xc002aabd40?, 0x1, 0xa?) test/e2e/framework/service/jig.go:222 > k8s.io/kubernetes/test/e2e/network.glob..func20.3() test/e2e/network/loadbalancer.go:1271 k8s.io/kubernetes/vendor/github.com/onsi/ginkgo/v2/internal.extractBodyFunction.func3({0x2d591ce, 0xc0004cfe00}) vendor/github.com/onsi/ginkgo/v2/internal/node.go:449 k8s.io/kubernetes/vendor/github.com/onsi/ginkgo/v2/internal.(*Suite).runNode.func2() vendor/github.com/onsi/ginkgo/v2/internal/suite.go:750 k8s.io/kubernetes/vendor/github.com/onsi/ginkgo/v2/internal.(*Suite).runNode vendor/github.com/onsi/ginkgo/v2/internal/suite.go:738 ------------------------------ ------------------------------ Progress Report for Ginkgo Process #11 Automatically polling progress: [sig-network] LoadBalancers ESIPP [Slow] should work for type=LoadBalancer (Spec Runtime: 7m20.672s) test/e2e/network/loadbalancer.go:1266 In [It] (Node Runtime: 7m20.02s) test/e2e/network/loadbalancer.go:1266 At [By Step] waiting for loadbalancer for service esipp-2671/external-local-lb (Step Runtime: 7m19.703s) test/e2e/framework/service/jig.go:260 Spec Goroutine goroutine 2087 [select] k8s.io/kubernetes/vendor/k8s.io/apimachinery/pkg/util/wait.WaitForWithContext({0x7fe0bc8, 0xc00013a000}, 0xc001dd11a0, 0x2fdb16a?) vendor/k8s.io/apimachinery/pkg/util/wait/wait.go:660 k8s.io/kubernetes/vendor/k8s.io/apimachinery/pkg/util/wait.poll({0x7fe0bc8, 0xc00013a000}, 0x28?, 0x2fd9d05?, 0x20?) vendor/k8s.io/apimachinery/pkg/util/wait/wait.go:596 k8s.io/kubernetes/vendor/k8s.io/apimachinery/pkg/util/wait.PollImmediateWithContext({0x7fe0bc8, 0xc00013a000}, 0xc001534f00?, 0xc0055adb78?, 0x262a967?) vendor/k8s.io/apimachinery/pkg/util/wait/wait.go:528 k8s.io/kubernetes/vendor/k8s.io/apimachinery/pkg/util/wait.PollImmediate(0xc0010da770?, 0x7fa7740?, 0xc000214bc0?) vendor/k8s.io/apimachinery/pkg/util/wait/wait.go:514 > k8s.io/kubernetes/test/e2e/framework/service.(*TestJig).waitForCondition(0xc0047adef0, 0x4?, {0x7600fe2, 0x14}, 0x7895b68) test/e2e/framework/service/jig.go:631 > k8s.io/kubernetes/test/e2e/framework/service.(*TestJig).WaitForLoadBalancer(0xc0047adef0, 0x41?) test/e2e/framework/service/jig.go:582 > k8s.io/kubernetes/test/e2e/framework/service.(*TestJig).CreateLoadBalancerService(0xc0047adef0, 0x6aba880?, 0xc0055ade28) test/e2e/framework/service/jig.go:261 > k8s.io/kubernetes/test/e2e/framework/service.(*TestJig).CreateOnlyLocalLoadBalancerService(0xc0047adef0, 0xc002aabd40?, 0x1, 0xa?) test/e2e/framework/service/jig.go:222 > k8s.io/kubernetes/test/e2e/network.glob..func20.3() test/e2e/network/loadbalancer.go:1271 k8s.io/kubernetes/vendor/github.com/onsi/ginkgo/v2/internal.extractBodyFunction.func3({0x2d591ce, 0xc0004cfe00}) vendor/github.com/onsi/ginkgo/v2/internal/node.go:449 k8s.io/kubernetes/vendor/github.com/onsi/ginkgo/v2/internal.(*Suite).runNode.func2() vendor/github.com/onsi/ginkgo/v2/internal/suite.go:750 k8s.io/kubernetes/vendor/github.com/onsi/ginkgo/v2/internal.(*Suite).runNode vendor/github.com/onsi/ginkgo/v2/internal/suite.go:738 ------------------------------ ------------------------------ Progress Report for Ginkgo Process #11 Automatically polling progress: [sig-network] LoadBalancers ESIPP [Slow] should work for type=LoadBalancer (Spec Runtime: 7m40.677s) test/e2e/network/loadbalancer.go:1266 In [It] (Node Runtime: 7m40.025s) test/e2e/network/loadbalancer.go:1266 At [By Step] waiting for loadbalancer for service esipp-2671/external-local-lb (Step Runtime: 7m39.707s) test/e2e/framework/service/jig.go:260 Spec Goroutine goroutine 2087 [select] k8s.io/kubernetes/vendor/k8s.io/apimachinery/pkg/util/wait.WaitForWithContext({0x7fe0bc8, 0xc00013a000}, 0xc001dd11a0, 0x2fdb16a?) vendor/k8s.io/apimachinery/pkg/util/wait/wait.go:660 k8s.io/kubernetes/vendor/k8s.io/apimachinery/pkg/util/wait.poll({0x7fe0bc8, 0xc00013a000}, 0x28?, 0x2fd9d05?, 0x20?) vendor/k8s.io/apimachinery/pkg/util/wait/wait.go:596 k8s.io/kubernetes/vendor/k8s.io/apimachinery/pkg/util/wait.PollImmediateWithContext({0x7fe0bc8, 0xc00013a000}, 0xc001534f00?, 0xc0055adb78?, 0x262a967?) vendor/k8s.io/apimachinery/pkg/util/wait/wait.go:528 k8s.io/kubernetes/vendor/k8s.io/apimachinery/pkg/util/wait.PollImmediate(0xc0010da770?, 0x7fa7740?, 0xc000214bc0?) vendor/k8s.io/apimachinery/pkg/util/wait/wait.go:514 > k8s.io/kubernetes/test/e2e/framework/service.(*TestJig).waitForCondition(0xc0047adef0, 0x4?, {0x7600fe2, 0x14}, 0x7895b68) test/e2e/framework/service/jig.go:631 > k8s.io/kubernetes/test/e2e/framework/service.(*TestJig).WaitForLoadBalancer(0xc0047adef0, 0x41?) test/e2e/framework/service/jig.go:582 > k8s.io/kubernetes/test/e2e/framework/service.(*TestJig).CreateLoadBalancerService(0xc0047adef0, 0x6aba880?, 0xc0055ade28) test/e2e/framework/service/jig.go:261 > k8s.io/kubernetes/test/e2e/framework/service.(*TestJig).CreateOnlyLocalLoadBalancerService(0xc0047adef0, 0xc002aabd40?, 0x1, 0xa?) test/e2e/framework/service/jig.go:222 > k8s.io/kubernetes/test/e2e/network.glob..func20.3() test/e2e/network/loadbalancer.go:1271 k8s.io/kubernetes/vendor/github.com/onsi/ginkgo/v2/internal.extractBodyFunction.func3({0x2d591ce, 0xc0004cfe00}) vendor/github.com/onsi/ginkgo/v2/internal/node.go:449 k8s.io/kubernetes/vendor/github.com/onsi/ginkgo/v2/internal.(*Suite).runNode.func2() vendor/github.com/onsi/ginkgo/v2/internal/suite.go:750 k8s.io/kubernetes/vendor/github.com/onsi/ginkgo/v2/internal.(*Suite).runNode vendor/github.com/onsi/ginkgo/v2/internal/suite.go:738 ------------------------------ ------------------------------ Progress Report for Ginkgo Process #11 Automatically polling progress: [sig-network] LoadBalancers ESIPP [Slow] should work for type=LoadBalancer (Spec Runtime: 8m0.679s) test/e2e/network/loadbalancer.go:1266 In [It] (Node Runtime: 8m0.028s) test/e2e/network/loadbalancer.go:1266 At [By Step] waiting for loadbalancer for service esipp-2671/external-local-lb (Step Runtime: 7m59.71s) test/e2e/framework/service/jig.go:260 Spec Goroutine goroutine 2087 [select] k8s.io/kubernetes/vendor/k8s.io/apimachinery/pkg/util/wait.WaitForWithContext({0x7fe0bc8, 0xc00013a000}, 0xc001dd11a0, 0x2fdb16a?) vendor/k8s.io/apimachinery/pkg/util/wait/wait.go:660 k8s.io/kubernetes/vendor/k8s.io/apimachinery/pkg/util/wait.poll({0x7fe0bc8, 0xc00013a000}, 0x28?, 0x2fd9d05?, 0x20?) vendor/k8s.io/apimachinery/pkg/util/wait/wait.go:596 k8s.io/kubernetes/vendor/k8s.io/apimachinery/pkg/util/wait.PollImmediateWithContext({0x7fe0bc8, 0xc00013a000}, 0xc001534f00?, 0xc0055adb78?, 0x262a967?) vendor/k8s.io/apimachinery/pkg/util/wait/wait.go:528 k8s.io/kubernetes/vendor/k8s.io/apimachinery/pkg/util/wait.PollImmediate(0xc0010da770?, 0x7fa7740?, 0xc000214bc0?) vendor/k8s.io/apimachinery/pkg/util/wait/wait.go:514 > k8s.io/kubernetes/test/e2e/framework/service.(*TestJig).waitForCondition(0xc0047adef0, 0x4?, {0x7600fe2, 0x14}, 0x7895b68) test/e2e/framework/service/jig.go:631 > k8s.io/kubernetes/test/e2e/framework/service.(*TestJig).WaitForLoadBalancer(0xc0047adef0, 0x41?) test/e2e/framework/service/jig.go:582 > k8s.io/kubernetes/test/e2e/framework/service.(*TestJig).CreateLoadBalancerService(0xc0047adef0, 0x6aba880?, 0xc0055ade28) test/e2e/framework/service/jig.go:261 > k8s.io/kubernetes/test/e2e/framework/service.(*TestJig).CreateOnlyLocalLoadBalancerService(0xc0047adef0, 0xc002aabd40?, 0x1, 0xa?) test/e2e/framework/service/jig.go:222 > k8s.io/kubernetes/test/e2e/network.glob..func20.3() test/e2e/network/loadbalancer.go:1271 k8s.io/kubernetes/vendor/github.com/onsi/ginkgo/v2/internal.extractBodyFunction.func3({0x2d591ce, 0xc0004cfe00}) vendor/github.com/onsi/ginkgo/v2/internal/node.go:449 k8s.io/kubernetes/vendor/github.com/onsi/ginkgo/v2/internal.(*Suite).runNode.func2() vendor/github.com/onsi/ginkgo/v2/internal/suite.go:750 k8s.io/kubernetes/vendor/github.com/onsi/ginkgo/v2/internal.(*Suite).runNode vendor/github.com/onsi/ginkgo/v2/internal/suite.go:738 ------------------------------ ------------------------------ Progress Report for Ginkgo Process #11 Automatically polling progress: [sig-network] LoadBalancers ESIPP [Slow] should work for type=LoadBalancer (Spec Runtime: 8m20.682s) test/e2e/network/loadbalancer.go:1266 In [It] (Node Runtime: 8m20.03s) test/e2e/network/loadbalancer.go:1266 At [By Step] waiting for loadbalancer for service esipp-2671/external-local-lb (Step Runtime: 8m19.712s) test/e2e/framework/service/jig.go:260 Spec Goroutine goroutine 2087 [select] k8s.io/kubernetes/vendor/golang.org/x/net/http2.(*ClientConn).RoundTrip(0xc0004ce180, 0xc00308d100) vendor/golang.org/x/net/http2/transport.go:1200 k8s.io/kubernetes/vendor/golang.org/x/net/http2.(*Transport).RoundTripOpt(0xc003f27900, 0xc00308d100, {0xe0?}) vendor/golang.org/x/net/http2/transport.go:519 k8s.io/kubernetes/vendor/golang.org/x/net/http2.(*Transport).RoundTrip(...) vendor/golang.org/x/net/http2/transport.go:480 k8s.io/kubernetes/vendor/golang.org/x/net/http2.noDialH2RoundTripper.RoundTrip({0xc001753b80?}, 0xc00308d100?) vendor/golang.org/x/net/http2/transport.go:3020 net/http.(*Transport).roundTrip(0xc001753b80, 0xc00308d100) /usr/local/go/src/net/http/transport.go:540 net/http.(*Transport).RoundTrip(0x6fe4b20?, 0xc003ef3a40?) /usr/local/go/src/net/http/roundtrip.go:17 k8s.io/kubernetes/vendor/k8s.io/client-go/transport.(*bearerAuthRoundTripper).RoundTrip(0xc0043ee360, 0xc00308d000) vendor/k8s.io/client-go/transport/round_trippers.go:317 k8s.io/kubernetes/vendor/k8s.io/client-go/transport.(*userAgentRoundTripper).RoundTrip(0xc0011b9fe0, 0xc00308cf00) vendor/k8s.io/client-go/transport/round_trippers.go:168 net/http.send(0xc00308cf00, {0x7fad100, 0xc0011b9fe0}, {0x74d54e0?, 0x1?, 0x0?}) /usr/local/go/src/net/http/client.go:251 net/http.(*Client).send(0xc0043ee390, 0xc00308cf00, {0x7f17692575b8?, 0x100?, 0x0?}) /usr/local/go/src/net/http/client.go:175 net/http.(*Client).do(0xc0043ee390, 0xc00308cf00) /usr/local/go/src/net/http/client.go:715 net/http.(*Client).Do(...) /usr/local/go/src/net/http/client.go:581 k8s.io/kubernetes/vendor/k8s.io/client-go/rest.(*Request).request(0xc00308cd00, {0x7fe0bc8, 0xc00013a008}, 0x0?) vendor/k8s.io/client-go/rest/request.go:964 k8s.io/kubernetes/vendor/k8s.io/client-go/rest.(*Request).Do(0xc00308cd00, {0x7fe0bc8, 0xc00013a008}) vendor/k8s.io/client-go/rest/request.go:1005 k8s.io/kubernetes/vendor/k8s.io/client-go/kubernetes/typed/core/v1.(*services).Get(0xc004283e60, {0x7fe0bc8, 0xc00013a008}, {0x75ed637, 0x11}, {{{0x0, 0x0}, {0x0, 0x0}}, {0x0, ...}}) vendor/k8s.io/client-go/kubernetes/typed/core/v1/service.go:79 > k8s.io/kubernetes/test/e2e/framework/service.(*TestJig).waitForCondition.func1() test/e2e/framework/service/jig.go:620 k8s.io/kubernetes/vendor/k8s.io/apimachinery/pkg/util/wait.ConditionFunc.WithContext.func1({0x2742871, 0x0}) vendor/k8s.io/apimachinery/pkg/util/wait/wait.go:222 k8s.io/kubernetes/vendor/k8s.io/apimachinery/pkg/util/wait.runConditionWithCrashProtectionWithContext({0x7fe0bc8?, 0xc00013a000?}, 0xc0046b81f8?) vendor/k8s.io/apimachinery/pkg/util/wait/wait.go:235 k8s.io/kubernetes/vendor/k8s.io/apimachinery/pkg/util/wait.WaitForWithContext({0x7fe0bc8, 0xc00013a000}, 0xc001dd11a0, 0x2fdb16a?) vendor/k8s.io/apimachinery/pkg/util/wait/wait.go:662 k8s.io/kubernetes/vendor/k8s.io/apimachinery/pkg/util/wait.poll({0x7fe0bc8, 0xc00013a000}, 0x28?, 0x2fd9d05?, 0x20?) vendor/k8s.io/apimachinery/pkg/util/wait/wait.go:596 k8s.io/kubernetes/vendor/k8s.io/apimachinery/pkg/util/wait.PollImmediateWithContext({0x7fe0bc8, 0xc00013a000}, 0xc001534f00?, 0xc0055adb78?, 0x262a967?) vendor/k8s.io/apimachinery/pkg/util/wait/wait.go:528 k8s.io/kubernetes/vendor/k8s.io/apimachinery/pkg/util/wait.PollImmediate(0xc0010da770?, 0x7fa7740?, 0xc000214bc0?) vendor/k8s.io/apimachinery/pkg/util/wait/wait.go:514 > k8s.io/kubernetes/test/e2e/framework/service.(*TestJig).waitForCondition(0xc0047adef0, 0x4?, {0x7600fe2, 0x14}, 0x7895b68) test/e2e/framework/service/jig.go:631 > k8s.io/kubernetes/test/e2e/framework/service.(*TestJig).WaitForLoadBalancer(0xc0047adef0, 0x41?) test/e2e/framework/service/jig.go:582 > k8s.io/kubernetes/test/e2e/framework/service.(*TestJig).CreateLoadBalancerService(0xc0047adef0, 0x6aba880?, 0xc0055ade28) test/e2e/framework/service/jig.go:261 > k8s.io/kubernetes/test/e2e/framework/service.(*TestJig).CreateOnlyLocalLoadBalancerService(0xc0047adef0, 0xc002aabd40?, 0x1, 0xa?) test/e2e/framework/service/jig.go:222 > k8s.io/kubernetes/test/e2e/network.glob..func20.3() test/e2e/network/loadbalancer.go:1271 k8s.io/kubernetes/vendor/github.com/onsi/ginkgo/v2/internal.extractBodyFunction.func3({0x2d591ce, 0xc0004cfe00}) vendor/github.com/onsi/ginkgo/v2/internal/node.go:449 k8s.io/kubernetes/vendor/github.com/onsi/ginkgo/v2/internal.(*Suite).runNode.func2() vendor/github.com/onsi/ginkgo/v2/internal/suite.go:750 k8s.io/kubernetes/vendor/github.com/onsi/ginkgo/v2/internal.(*Suite).runNode vendor/github.com/onsi/ginkgo/v2/internal/suite.go:738 ------------------------------ ------------------------------ Progress Report for Ginkgo Process #11 Automatically polling progress: [sig-network] LoadBalancers ESIPP [Slow] should work for type=LoadBalancer (Spec Runtime: 8m40.685s) test/e2e/network/loadbalancer.go:1266 In [It] (Node Runtime: 8m40.033s) test/e2e/network/loadbalancer.go:1266 At [By Step] waiting for loadbalancer for service esipp-2671/external-local-lb (Step Runtime: 8m39.715s) test/e2e/framework/service/jig.go:260 Spec Goroutine goroutine 2087 [select] k8s.io/kubernetes/vendor/golang.org/x/net/http2.(*ClientConn).RoundTrip(0xc0004ce180, 0xc00308d100) vendor/golang.org/x/net/http2/transport.go:1200 k8s.io/kubernetes/vendor/golang.org/x/net/http2.(*Transport).RoundTripOpt(0xc003f27900, 0xc00308d100, {0xe0?}) vendor/golang.org/x/net/http2/transport.go:519 k8s.io/kubernetes/vendor/golang.org/x/net/http2.(*Transport).RoundTrip(...) vendor/golang.org/x/net/http2/transport.go:480 k8s.io/kubernetes/vendor/golang.org/x/net/http2.noDialH2RoundTripper.RoundTrip({0xc001753b80?}, 0xc00308d100?) vendor/golang.org/x/net/http2/transport.go:3020 net/http.(*Transport).roundTrip(0xc001753b80, 0xc00308d100) /usr/local/go/src/net/http/transport.go:540 net/http.(*Transport).RoundTrip(0x6fe4b20?, 0xc003ef3a40?) /usr/local/go/src/net/http/roundtrip.go:17 k8s.io/kubernetes/vendor/k8s.io/client-go/transport.(*bearerAuthRoundTripper).RoundTrip(0xc0043ee360, 0xc00308d000) vendor/k8s.io/client-go/transport/round_trippers.go:317 k8s.io/kubernetes/vendor/k8s.io/client-go/transport.(*userAgentRoundTripper).RoundTrip(0xc0011b9fe0, 0xc00308cf00) vendor/k8s.io/client-go/transport/round_trippers.go:168 net/http.send(0xc00308cf00, {0x7fad100, 0xc0011b9fe0}, {0x74d54e0?, 0x1?, 0x0?}) /usr/local/go/src/net/http/client.go:251 net/http.(*Client).send(0xc0043ee390, 0xc00308cf00, {0x7f17692575b8?, 0x100?, 0x0?}) /usr/local/go/src/net/http/client.go:175 net/http.(*Client).do(0xc0043ee390, 0xc00308cf00) /usr/local/go/src/net/http/client.go:715 net/http.(*Client).Do(...) /usr/local/go/src/net/http/client.go:581 k8s.io/kubernetes/vendor/k8s.io/client-go/rest.(*Request).request(0xc00308cd00, {0x7fe0bc8, 0xc00013a008}, 0x0?) vendor/k8s.io/client-go/rest/request.go:964 k8s.io/kubernetes/vendor/k8s.io/client-go/rest.(*Request).Do(0xc00308cd00, {0x7fe0bc8, 0xc00013a008}) vendor/k8s.io/client-go/rest/request.go:1005 k8s.io/kubernetes/vendor/k8s.io/client-go/kubernetes/typed/core/v1.(*services).Get(0xc004283e60, {0x7fe0bc8, 0xc00013a008}, {0x75ed637, 0x11}, {{{0x0, 0x0}, {0x0, 0x0}}, {0x0, ...}}) vendor/k8s.io/client-go/kubernetes/typed/core/v1/service.go:79 > k8s.io/kubernetes/test/e2e/framework/service.(*TestJig).waitForCondition.func1() test/e2e/framework/service/jig.go:620 k8s.io/kubernetes/vendor/k8s.io/apimachinery/pkg/util/wait.ConditionFunc.WithContext.func1({0x2742871, 0x0}) vendor/k8s.io/apimachinery/pkg/util/wait/wait.go:222 k8s.io/kubernetes/vendor/k8s.io/apimachinery/pkg/util/wait.runConditionWithCrashProtectionWithContext({0x7fe0bc8?, 0xc00013a000?}, 0xc0046b81f8?) vendor/k8s.io/apimachinery/pkg/util/wait/wait.go:235 k8s.io/kubernetes/vendor/k8s.io/apimachinery/pkg/util/wait.WaitForWithContext({0x7fe0bc8, 0xc00013a000}, 0xc001dd11a0, 0x2fdb16a?) vendor/k8s.io/apimachinery/pkg/util/wait/wait.go:662 k8s.io/kubernetes/vendor/k8s.io/apimachinery/pkg/util/wait.poll({0x7fe0bc8, 0xc00013a000}, 0x28?, 0x2fd9d05?, 0x20?) vendor/k8s.io/apimachinery/pkg/util/wait/wait.go:596 k8s.io/kubernetes/vendor/k8s.io/apimachinery/pkg/util/wait.PollImmediateWithContext({0x7fe0bc8, 0xc00013a000}, 0xc001534f00?, 0xc0055adb78?, 0x262a967?) vendor/k8s.io/apimachinery/pkg/util/wait/wait.go:528 k8s.io/kubernetes/vendor/k8s.io/apimachinery/pkg/util/wait.PollImmediate(0xc0010da770?, 0x7fa7740?, 0xc000214bc0?) vendor/k8s.io/apimachinery/pkg/util/wait/wait.go:514 > k8s.io/kubernetes/test/e2e/framework/service.(*TestJig).waitForCondition(0xc0047adef0, 0x4?, {0x7600fe2, 0x14}, 0x7895b68) test/e2e/framework/service/jig.go:631 > k8s.io/kubernetes/test/e2e/framework/service.(*TestJig).WaitForLoadBalancer(0xc0047adef0, 0x41?) test/e2e/framework/service/jig.go:582 > k8s.io/kubernetes/test/e2e/framework/service.(*TestJig).CreateLoadBalancerService(0xc0047adef0, 0x6aba880?, 0xc0055ade28) test/e2e/framework/service/jig.go:261 > k8s.io/kubernetes/test/e2e/framework/service.(*TestJig).CreateOnlyLocalLoadBalancerService(0xc0047adef0, 0xc002aabd40?, 0x1, 0xa?) test/e2e/framework/service/jig.go:222 > k8s.io/kubernetes/test/e2e/network.glob..func20.3() test/e2e/network/loadbalancer.go:1271 k8s.io/kubernetes/vendor/github.com/onsi/ginkgo/v2/internal.extractBodyFunction.func3({0x2d591ce, 0xc0004cfe00}) vendor/github.com/onsi/ginkgo/v2/internal/node.go:449 k8s.io/kubernetes/vendor/github.com/onsi/ginkgo/v2/internal.(*Suite).runNode.func2() vendor/github.com/onsi/ginkgo/v2/internal/suite.go:750 k8s.io/kubernetes/vendor/github.com/onsi/ginkgo/v2/internal.(*Suite).runNode vendor/github.com/onsi/ginkgo/v2/internal/suite.go:738 ------------------------------ ------------------------------ Progress Report for Ginkgo Process #11 Automatically polling progress: [sig-network] LoadBalancers ESIPP [Slow] should work for type=LoadBalancer (Spec Runtime: 9m0.688s) test/e2e/network/loadbalancer.go:1266 In [It] (Node Runtime: 9m0.036s) test/e2e/network/loadbalancer.go:1266 At [By Step] waiting for loadbalancer for service esipp-2671/external-local-lb (Step Runtime: 8m59.718s) test/e2e/framework/service/jig.go:260 Spec Goroutine goroutine 2087 [select] k8s.io/kubernetes/vendor/golang.org/x/net/http2.(*ClientConn).RoundTrip(0xc0004ce180, 0xc00308d100) vendor/golang.org/x/net/http2/transport.go:1200 k8s.io/kubernetes/vendor/golang.org/x/net/http2.(*Transport).RoundTripOpt(0xc003f27900, 0xc00308d100, {0xe0?}) vendor/golang.org/x/net/http2/transport.go:519 k8s.io/kubernetes/vendor/golang.org/x/net/http2.(*Transport).RoundTrip(...) vendor/golang.org/x/net/http2/transport.go:480 k8s.io/kubernetes/vendor/golang.org/x/net/http2.noDialH2RoundTripper.RoundTrip({0xc001753b80?}, 0xc00308d100?) vendor/golang.org/x/net/http2/transport.go:3020 net/http.(*Transport).roundTrip(0xc001753b80, 0xc00308d100) /usr/local/go/src/net/http/transport.go:540 net/http.(*Transport).RoundTrip(0x6fe4b20?, 0xc003ef3a40?) /usr/local/go/src/net/http/roundtrip.go:17 k8s.io/kubernetes/vendor/k8s.io/client-go/transport.(*bearerAuthRoundTripper).RoundTrip(0xc0043ee360, 0xc00308d000) vendor/k8s.io/client-go/transport/round_trippers.go:317 k8s.io/kubernetes/vendor/k8s.io/client-go/transport.(*userAgentRoundTripper).RoundTrip(0xc0011b9fe0, 0xc00308cf00) vendor/k8s.io/client-go/transport/round_trippers.go:168 net/http.send(0xc00308cf00, {0x7fad100, 0xc0011b9fe0}, {0x74d54e0?, 0x1?, 0x0?}) /usr/local/go/src/net/http/client.go:251 net/http.(*Client).send(0xc0043ee390, 0xc00308cf00, {0x7f17692575b8?, 0x100?, 0x0?}) /usr/local/go/src/net/http/client.go:175 net/http.(*Client).do(0xc0043ee390, 0xc00308cf00) /usr/local/go/src/net/http/client.go:715 net/http.(*Client).Do(...) /usr/local/go/src/net/http/client.go:581 k8s.io/kubernetes/vendor/k8s.io/client-go/rest.(*Request).request(0xc00308cd00, {0x7fe0bc8, 0xc00013a008}, 0x0?) vendor/k8s.io/client-go/rest/request.go:964 k8s.io/kubernetes/vendor/k8s.io/client-go/rest.(*Request).Do(0xc00308cd00, {0x7fe0bc8, 0xc00013a008}) vendor/k8s.io/client-go/rest/request.go:1005 k8s.io/kubernetes/vendor/k8s.io/client-go/kubernetes/typed/core/v1.(*services).Get(0xc004283e60, {0x7fe0bc8, 0xc00013a008}, {0x75ed637, 0x11}, {{{0x0, 0x0}, {0x0, 0x0}}, {0x0, ...}}) vendor/k8s.io/client-go/kubernetes/typed/core/v1/service.go:79 > k8s.io/kubernetes/test/e2e/framework/service.(*TestJig).waitForCondition.func1() test/e2e/framework/service/jig.go:620 k8s.io/kubernetes/vendor/k8s.io/apimachinery/pkg/util/wait.ConditionFunc.WithContext.func1({0x2742871, 0x0}) vendor/k8s.io/apimachinery/pkg/util/wait/wait.go:222 k8s.io/kubernetes/vendor/k8s.io/apimachinery/pkg/util/wait.runConditionWithCrashProtectionWithContext({0x7fe0bc8?, 0xc00013a000?}, 0xc0046b81f8?) vendor/k8s.io/apimachinery/pkg/util/wait/wait.go:235 k8s.io/kubernetes/vendor/k8s.io/apimachinery/pkg/util/wait.WaitForWithContext({0x7fe0bc8, 0xc00013a000}, 0xc001dd11a0, 0x2fdb16a?) vendor/k8s.io/apimachinery/pkg/util/wait/wait.go:662 k8s.io/kubernetes/vendor/k8s.io/apimachinery/pkg/util/wait.poll({0x7fe0bc8, 0xc00013a000}, 0x28?, 0x2fd9d05?, 0x20?) vendor/k8s.io/apimachinery/pkg/util/wait/wait.go:596 k8s.io/kubernetes/vendor/k8s.io/apimachinery/pkg/util/wait.PollImmediateWithContext({0x7fe0bc8, 0xc00013a000}, 0xc001534f00?, 0xc0055adb78?, 0x262a967?) vendor/k8s.io/apimachinery/pkg/util/wait/wait.go:528 k8s.io/kubernetes/vendor/k8s.io/apimachinery/pkg/util/wait.PollImmediate(0xc0010da770?, 0x7fa7740?, 0xc000214bc0?) vendor/k8s.io/apimachinery/pkg/util/wait/wait.go:514 > k8s.io/kubernetes/test/e2e/framework/service.(*TestJig).waitForCondition(0xc0047adef0, 0x4?, {0x7600fe2, 0x14}, 0x7895b68) test/e2e/framework/service/jig.go:631 > k8s.io/kubernetes/test/e2e/framework/service.(*TestJig).WaitForLoadBalancer(0xc0047adef0, 0x41?) test/e2e/framework/service/jig.go:582 > k8s.io/kubernetes/test/e2e/framework/service.(*TestJig).CreateLoadBalancerService(0xc0047adef0, 0x6aba880?, 0xc0055ade28) test/e2e/framework/service/jig.go:261 > k8s.io/kubernetes/test/e2e/framework/service.(*TestJig).CreateOnlyLocalLoadBalancerService(0xc0047adef0, 0xc002aabd40?, 0x1, 0xa?) test/e2e/framework/service/jig.go:222 > k8s.io/kubernetes/test/e2e/network.glob..func20.3() test/e2e/network/loadbalancer.go:1271 k8s.io/kubernetes/vendor/github.com/onsi/ginkgo/v2/internal.extractBodyFunction.func3({0x2d591ce, 0xc0004cfe00}) vendor/github.com/onsi/ginkgo/v2/internal/node.go:449 k8s.io/kubernetes/vendor/github.com/onsi/ginkgo/v2/internal.(*Suite).runNode.func2() vendor/github.com/onsi/ginkgo/v2/internal/suite.go:750 k8s.io/kubernetes/vendor/github.com/onsi/ginkgo/v2/internal.(*Suite).runNode vendor/github.com/onsi/ginkgo/v2/internal/suite.go:738 ------------------------------ Nov 26 05:26:41.493: INFO: Retrying .... error trying to get Service external-local-lb: Get "https://35.227.179.144/api/v1/namespaces/esipp-2671/services/external-local-lb": stream error: stream ID 803; INTERNAL_ERROR; received from peer ------------------------------ Progress Report for Ginkgo Process #11 Automatically polling progress: [sig-network] LoadBalancers ESIPP [Slow] should work for type=LoadBalancer (Spec Runtime: 9m20.69s) test/e2e/network/loadbalancer.go:1266 In [It] (Node Runtime: 9m20.038s) test/e2e/network/loadbalancer.go:1266 At [By Step] waiting for loadbalancer for service esipp-2671/external-local-lb (Step Runtime: 9m19.72s) test/e2e/framework/service/jig.go:260 Spec Goroutine goroutine 2087 [select] k8s.io/kubernetes/vendor/golang.org/x/net/http2.(*ClientConn).RoundTrip(0xc0004ce180, 0xc00308d600) vendor/golang.org/x/net/http2/transport.go:1200 k8s.io/kubernetes/vendor/golang.org/x/net/http2.(*Transport).RoundTripOpt(0xc003f27900, 0xc00308d600, {0xe0?}) vendor/golang.org/x/net/http2/transport.go:519 k8s.io/kubernetes/vendor/golang.org/x/net/http2.(*Transport).RoundTrip(...) vendor/golang.org/x/net/http2/transport.go:480 k8s.io/kubernetes/vendor/golang.org/x/net/http2.noDialH2RoundTripper.RoundTrip({0xc001753b80?}, 0xc00308d600?) vendor/golang.org/x/net/http2/transport.go:3020 net/http.(*Transport).roundTrip(0xc001753b80, 0xc00308d600) /usr/local/go/src/net/http/transport.go:540 net/http.(*Transport).RoundTrip(0x6fe4b20?, 0xc003ef3c80?) /usr/local/go/src/net/http/roundtrip.go:17 k8s.io/kubernetes/vendor/k8s.io/client-go/transport.(*bearerAuthRoundTripper).RoundTrip(0xc0043ee360, 0xc00308d500) vendor/k8s.io/client-go/transport/round_trippers.go:317 k8s.io/kubernetes/vendor/k8s.io/client-go/transport.(*userAgentRoundTripper).RoundTrip(0xc0011b9fe0, 0xc00308d400) vendor/k8s.io/client-go/transport/round_trippers.go:168 net/http.send(0xc00308d400, {0x7fad100, 0xc0011b9fe0}, {0x74d54e0?, 0x1?, 0x0?}) /usr/local/go/src/net/http/client.go:251 net/http.(*Client).send(0xc0043ee390, 0xc00308d400, {0x7f17692575b8?, 0x100?, 0x0?}) /usr/local/go/src/net/http/client.go:175 net/http.(*Client).do(0xc0043ee390, 0xc00308d400) /usr/local/go/src/net/http/client.go:715 net/http.(*Client).Do(...) /usr/local/go/src/net/http/client.go:581 k8s.io/kubernetes/vendor/k8s.io/client-go/rest.(*Request).request(0xc00308d200, {0x7fe0bc8, 0xc00013a008}, 0x0?) vendor/k8s.io/client-go/rest/request.go:964 k8s.io/kubernetes/vendor/k8s.io/client-go/rest.(*Request).Do(0xc00308d200, {0x7fe0bc8, 0xc00013a008}) vendor/k8s.io/client-go/rest/request.go:1005 k8s.io/kubernetes/vendor/k8s.io/client-go/kubernetes/typed/core/v1.(*services).Get(0xc004283f00, {0x7fe0bc8, 0xc00013a008}, {0x75ed637, 0x11}, {{{0x0, 0x0}, {0x0, 0x0}}, {0x0, ...}}) vendor/k8s.io/client-go/kubernetes/typed/core/v1/service.go:79 > k8s.io/kubernetes/test/e2e/framework/service.(*TestJig).waitForCondition.func1() test/e2e/framework/service/jig.go:620 k8s.io/kubernetes/vendor/k8s.io/apimachinery/pkg/util/wait.ConditionFunc.WithContext.func1({0x2742871, 0x0}) vendor/k8s.io/apimachinery/pkg/util/wait/wait.go:222 k8s.io/kubernetes/vendor/k8s.io/apimachinery/pkg/util/wait.runConditionWithCrashProtectionWithContext({0x7fe0bc8?, 0xc00013a000?}, 0xc0046b81f8?) vendor/k8s.io/apimachinery/pkg/util/wait/wait.go:235 k8s.io/kubernetes/vendor/k8s.io/apimachinery/pkg/util/wait.WaitForWithContext({0x7fe0bc8, 0xc00013a000}, 0xc001dd11a0, 0x2fdb16a?) vendor/k8s.io/apimachinery/pkg/util/wait/wait.go:662 k8s.io/kubernetes/vendor/k8s.io/apimachinery/pkg/util/wait.poll({0x7fe0bc8, 0xc00013a000}, 0x28?, 0x2fd9d05?, 0x20?) vendor/k8s.io/apimachinery/pkg/util/wait/wait.go:596 k8s.io/kubernetes/vendor/k8s.io/apimachinery/pkg/util/wait.PollImmediateWithContext({0x7fe0bc8, 0xc00013a000}, 0xc001534f00?, 0xc0055adb78?, 0x262a967?) vendor/k8s.io/apimachinery/pkg/util/wait/wait.go:528 k8s.io/kubernetes/vendor/k8s.io/apimachinery/pkg/util/wait.PollImmediate(0xc0010da770?, 0x7fa7740?, 0xc000214bc0?) vendor/k8s.io/apimachinery/pkg/util/wait/wait.go:514 > k8s.io/kubernetes/test/e2e/framework/service.(*TestJig).waitForCondition(0xc0047adef0, 0x4?, {0x7600fe2, 0x14}, 0x7895b68) test/e2e/framework/service/jig.go:631 > k8s.io/kubernetes/test/e2e/framework/service.(*TestJig).WaitForLoadBalancer(0xc0047adef0, 0x41?) test/e2e/framework/service/jig.go:582 > k8s.io/kubernetes/test/e2e/framework/service.(*TestJig).CreateLoadBalancerService(0xc0047adef0, 0x6aba880?, 0xc0055ade28) test/e2e/framework/service/jig.go:261 > k8s.io/kubernetes/test/e2e/framework/service.(*TestJig).CreateOnlyLocalLoadBalancerService(0xc0047adef0, 0xc002aabd40?, 0x1, 0xa?) test/e2e/framework/service/jig.go:222 > k8s.io/kubernetes/test/e2e/network.glob..func20.3() test/e2e/network/loadbalancer.go:1271 k8s.io/kubernetes/vendor/github.com/onsi/ginkgo/v2/internal.extractBodyFunction.func3({0x2d591ce, 0xc0004cfe00}) vendor/github.com/onsi/ginkgo/v2/internal/node.go:449 k8s.io/kubernetes/vendor/github.com/onsi/ginkgo/v2/internal.(*Suite).runNode.func2() vendor/github.com/onsi/ginkgo/v2/internal/suite.go:750 k8s.io/kubernetes/vendor/github.com/onsi/ginkgo/v2/internal.(*Suite).runNode vendor/github.com/onsi/ginkgo/v2/internal/suite.go:738 ------------------------------ ------------------------------ Progress Report for Ginkgo Process #11 Automatically polling progress: [sig-network] LoadBalancers ESIPP [Slow] should work for type=LoadBalancer (Spec Runtime: 9m40.693s) test/e2e/network/loadbalancer.go:1266 In [It] (Node Runtime: 9m40.041s) test/e2e/network/loadbalancer.go:1266 At [By Step] waiting for loadbalancer for service esipp-2671/external-local-lb (Step Runtime: 9m39.724s) test/e2e/framework/service/jig.go:260 Spec Goroutine goroutine 2087 [select] k8s.io/kubernetes/vendor/golang.org/x/net/http2.(*ClientConn).RoundTrip(0xc0004ce180, 0xc00308d600) vendor/golang.org/x/net/http2/transport.go:1200 k8s.io/kubernetes/vendor/golang.org/x/net/http2.(*Transport).RoundTripOpt(0xc003f27900, 0xc00308d600, {0xe0?}) vendor/golang.org/x/net/http2/transport.go:519 k8s.io/kubernetes/vendor/golang.org/x/net/http2.(*Transport).RoundTrip(...) vendor/golang.org/x/net/http2/transport.go:480 k8s.io/kubernetes/vendor/golang.org/x/net/http2.noDialH2RoundTripper.RoundTrip({0xc001753b80?}, 0xc00308d600?) vendor/golang.org/x/net/http2/transport.go:3020 net/http.(*Transport).roundTrip(0xc001753b80, 0xc00308d600) /usr/local/go/src/net/http/transport.go:540 net/http.(*Transport).RoundTrip(0x6fe4b20?, 0xc003ef3c80?) /usr/local/go/src/net/http/roundtrip.go:17 k8s.io/kubernetes/vendor/k8s.io/client-go/transport.(*bearerAuthRoundTripper).RoundTrip(0xc0043ee360, 0xc00308d500) vendor/k8s.io/client-go/transport/round_trippers.go:317 k8s.io/kubernetes/vendor/k8s.io/client-go/transport.(*userAgentRoundTripper).RoundTrip(0xc0011b9fe0, 0xc00308d400) vendor/k8s.io/client-go/transport/round_trippers.go:168 net/http.send(0xc00308d400, {0x7fad100, 0xc0011b9fe0}, {0x74d54e0?, 0x1?, 0x0?}) /usr/local/go/src/net/http/client.go:251 net/http.(*Client).send(0xc0043ee390, 0xc00308d400, {0x7f17692575b8?, 0x100?, 0x0?}) /usr/local/go/src/net/http/client.go:175 net/http.(*Client).do(0xc0043ee390, 0xc00308d400) /usr/local/go/src/net/http/client.go:715 net/http.(*Client).Do(...) /usr/local/go/src/net/http/client.go:581 k8s.io/kubernetes/vendor/k8s.io/client-go/rest.(*Request).request(0xc00308d200, {0x7fe0bc8, 0xc00013a008}, 0x0?) vendor/k8s.io/client-go/rest/request.go:964 k8s.io/kubernetes/vendor/k8s.io/client-go/rest.(*Request).Do(0xc00308d200, {0x7fe0bc8, 0xc00013a008}) vendor/k8s.io/client-go/rest/request.go:1005 k8s.io/kubernetes/vendor/k8s.io/client-go/kubernetes/typed/core/v1.(*services).Get(0xc004283f00, {0x7fe0bc8, 0xc00013a008}, {0x75ed637, 0x11}, {{{0x0, 0x0}, {0x0, 0x0}}, {0x0, ...}}) vendor/k8s.io/client-go/kubernetes/typed/core/v1/service.go:79 > k8s.io/kubernetes/test/e2e/framework/service.(*TestJig).waitForCondition.func1() test/e2e/framework/service/jig.go:620 k8s.io/kubernetes/vendor/k8s.io/apimachinery/pkg/util/wait.ConditionFunc.WithContext.func1({0x2742871, 0x0}) vendor/k8s.io/apimachinery/pkg/util/wait/wait.go:222 k8s.io/kubernetes/vendor/k8s.io/apimachinery/pkg/util/wait.runConditionWithCrashProtectionWithContext({0x7fe0bc8?, 0xc00013a000?}, 0xc0046b81f8?) vendor/k8s.io/apimachinery/pkg/util/wait/wait.go:235 k8s.io/kubernetes/vendor/k8s.io/apimachinery/pkg/util/wait.WaitForWithContext({0x7fe0bc8, 0xc00013a000}, 0xc001dd11a0, 0x2fdb16a?) vendor/k8s.io/apimachinery/pkg/util/wait/wait.go:662 k8s.io/kubernetes/vendor/k8s.io/apimachinery/pkg/util/wait.poll({0x7fe0bc8, 0xc00013a000}, 0x28?, 0x2fd9d05?, 0x20?) vendor/k8s.io/apimachinery/pkg/util/wait/wait.go:596 k8s.io/kubernetes/vendor/k8s.io/apimachinery/pkg/util/wait.PollImmediateWithContext({0x7fe0bc8, 0xc00013a000}, 0xc001534f00?, 0xc0055adb78?, 0x262a967?) vendor/k8s.io/apimachinery/pkg/util/wait/wait.go:528 k8s.io/kubernetes/vendor/k8s.io/apimachinery/pkg/util/wait.PollImmediate(0xc0010da770?, 0x7fa7740?, 0xc000214bc0?) vendor/k8s.io/apimachinery/pkg/util/wait/wait.go:514 > k8s.io/kubernetes/test/e2e/framework/service.(*TestJig).waitForCondition(0xc0047adef0, 0x4?, {0x7600fe2, 0x14}, 0x7895b68) test/e2e/framework/service/jig.go:631 > k8s.io/kubernetes/test/e2e/framework/service.(*TestJig).WaitForLoadBalancer(0xc0047adef0, 0x41?) test/e2e/framework/service/jig.go:582 > k8s.io/kubernetes/test/e2e/framework/service.(*TestJig).CreateLoadBalancerService(0xc0047adef0, 0x6aba880?, 0xc0055ade28) test/e2e/framework/service/jig.go:261 > k8s.io/kubernetes/test/e2e/framework/service.(*TestJig).CreateOnlyLocalLoadBalancerService(0xc0047adef0, 0xc002aabd40?, 0x1, 0xa?) test/e2e/framework/service/jig.go:222 > k8s.io/kubernetes/test/e2e/network.glob..func20.3() test/e2e/network/loadbalancer.go:1271 k8s.io/kubernetes/vendor/github.com/onsi/ginkgo/v2/internal.extractBodyFunction.func3({0x2d591ce, 0xc0004cfe00}) vendor/github.com/onsi/ginkgo/v2/internal/node.go:449 k8s.io/kubernetes/vendor/github.com/onsi/ginkgo/v2/internal.(*Suite).runNode.func2() vendor/github.com/onsi/ginkgo/v2/internal/suite.go:750 k8s.io/kubernetes/vendor/github.com/onsi/ginkgo/v2/internal.(*Suite).runNode vendor/github.com/onsi/ginkgo/v2/internal/suite.go:738 ------------------------------ ------------------------------ Progress Report for Ginkgo Process #11 Automatically polling progress: [sig-network] LoadBalancers ESIPP [Slow] should work for type=LoadBalancer (Spec Runtime: 10m0.696s) test/e2e/network/loadbalancer.go:1266 In [It] (Node Runtime: 10m0.044s) test/e2e/network/loadbalancer.go:1266 At [By Step] waiting for loadbalancer for service esipp-2671/external-local-lb (Step Runtime: 9m59.726s) test/e2e/framework/service/jig.go:260 Spec Goroutine goroutine 2087 [select, 2 minutes] k8s.io/kubernetes/vendor/golang.org/x/net/http2.(*ClientConn).RoundTrip(0xc0004ce180, 0xc00308d600) vendor/golang.org/x/net/http2/transport.go:1200 k8s.io/kubernetes/vendor/golang.org/x/net/http2.(*Transport).RoundTripOpt(0xc003f27900, 0xc00308d600, {0xe0?}) vendor/golang.org/x/net/http2/transport.go:519 k8s.io/kubernetes/vendor/golang.org/x/net/http2.(*Transport).RoundTrip(...) vendor/golang.org/x/net/http2/transport.go:480 k8s.io/kubernetes/vendor/golang.org/x/net/http2.noDialH2RoundTripper.RoundTrip({0xc001753b80?}, 0xc00308d600?) vendor/golang.org/x/net/http2/transport.go:3020 net/http.(*Transport).roundTrip(0xc001753b80, 0xc00308d600) /usr/local/go/src/net/http/transport.go:540 net/http.(*Transport).RoundTrip(0x6fe4b20?, 0xc003ef3c80?) /usr/local/go/src/net/http/roundtrip.go:17 k8s.io/kubernetes/vendor/k8s.io/client-go/transport.(*bearerAuthRoundTripper).RoundTrip(0xc0043ee360, 0xc00308d500) vendor/k8s.io/client-go/transport/round_trippers.go:317 k8s.io/kubernetes/vendor/k8s.io/client-go/transport.(*userAgentRoundTripper).RoundTrip(0xc0011b9fe0, 0xc00308d400) vendor/k8s.io/client-go/transport/round_trippers.go:168 net/http.send(0xc00308d400, {0x7fad100, 0xc0011b9fe0}, {0x74d54e0?, 0x1?, 0x0?}) /usr/local/go/src/net/http/client.go:251 net/http.(*Client).send(0xc0043ee390, 0xc00308d400, {0x7f17692575b8?, 0x100?, 0x0?}) /usr/local/go/src/net/http/client.go:175 net/http.(*Client).do(0xc0043ee390, 0xc00308d400) /usr/local/go/src/net/http/client.go:715 net/http.(*Client).Do(...) /usr/local/go/src/net/http/client.go:581 k8s.io/kubernetes/vendor/k8s.io/client-go/rest.(*Request).request(0xc00308d200, {0x7fe0bc8, 0xc00013a008}, 0x0?) vendor/k8s.io/client-go/rest/request.go:964 k8s.io/kubernetes/vendor/k8s.io/client-go/rest.(*Request).Do(0xc00308d200, {0x7fe0bc8, 0xc00013a008}) vendor/k8s.io/client-go/rest/request.go:1005 k8s.io/kubernetes/vendor/k8s.io/client-go/kubernetes/typed/core/v1.(*services).Get(0xc004283f00, {0x7fe0bc8, 0xc00013a008}, {0x75ed637, 0x11}, {{{0x0, 0x0}, {0x0, 0x0}}, {0x0, ...}}) vendor/k8s.io/client-go/kubernetes/typed/core/v1/service.go:79 > k8s.io/kubernetes/test/e2e/framework/service.(*TestJig).waitForCondition.func1() test/e2e/framework/service/jig.go:620 k8s.io/kubernetes/vendor/k8s.io/apimachinery/pkg/util/wait.ConditionFunc.WithContext.func1({0x2742871, 0x0}) vendor/k8s.io/apimachinery/pkg/util/wait/wait.go:222 k8s.io/kubernetes/vendor/k8s.io/apimachinery/pkg/util/wait.runConditionWithCrashProtectionWithContext({0x7fe0bc8?, 0xc00013a000?}, 0xc0046b81f8?) vendor/k8s.io/apimachinery/pkg/util/wait/wait.go:235 k8s.io/kubernetes/vendor/k8s.io/apimachinery/pkg/util/wait.WaitForWithContext({0x7fe0bc8, 0xc00013a000}, 0xc001dd11a0, 0x2fdb16a?) vendor/k8s.io/apimachinery/pkg/util/wait/wait.go:662 k8s.io/kubernetes/vendor/k8s.io/apimachinery/pkg/util/wait.poll({0x7fe0bc8, 0xc00013a000}, 0x28?, 0x2fd9d05?, 0x20?) vendor/k8s.io/apimachinery/pkg/util/wait/wait.go:596 k8s.io/kubernetes/vendor/k8s.io/apimachinery/pkg/util/wait.PollImmediateWithContext({0x7fe0bc8, 0xc00013a000}, 0xc001534f00?, 0xc0055adb78?, 0x262a967?) vendor/k8s.io/apimachinery/pkg/util/wait/wait.go:528 k8s.io/kubernetes/vendor/k8s.io/apimachinery/pkg/util/wait.PollImmediate(0xc0010da770?, 0x7fa7740?, 0xc000214bc0?) vendor/k8s.io/apimachinery/pkg/util/wait/wait.go:514 > k8s.io/kubernetes/test/e2e/framework/service.(*TestJig).waitForCondition(0xc0047adef0, 0x4?, {0x7600fe2, 0x14}, 0x7895b68) test/e2e/framework/service/jig.go:631 > k8s.io/kubernetes/test/e2e/framework/service.(*TestJig).WaitForLoadBalancer(0xc0047adef0, 0x41?) test/e2e/framework/service/jig.go:582 > k8s.io/kubernetes/test/e2e/framework/service.(*TestJig).CreateLoadBalancerService(0xc0047adef0, 0x6aba880?, 0xc0055ade28) test/e2e/framework/service/jig.go:261 > k8s.io/kubernetes/test/e2e/framework/service.(*TestJig).CreateOnlyLocalLoadBalancerService(0xc0047adef0, 0xc002aabd40?, 0x1, 0xa?) test/e2e/framework/service/jig.go:222 > k8s.io/kubernetes/test/e2e/network.glob..func20.3() test/e2e/network/loadbalancer.go:1271 k8s.io/kubernetes/vendor/github.com/onsi/ginkgo/v2/internal.extractBodyFunction.func3({0x2d591ce, 0xc0004cfe00}) vendor/github.com/onsi/ginkgo/v2/internal/node.go:449 k8s.io/kubernetes/vendor/github.com/onsi/ginkgo/v2/internal.(*Suite).runNode.func2() vendor/github.com/onsi/ginkgo/v2/internal/suite.go:750 k8s.io/kubernetes/vendor/github.com/onsi/ginkgo/v2/internal.(*Suite).runNode vendor/github.com/onsi/ginkgo/v2/internal/suite.go:738 ------------------------------ ------------------------------ Progress Report for Ginkgo Process #11 Automatically polling progress: [sig-network] LoadBalancers ESIPP [Slow] should work for type=LoadBalancer (Spec Runtime: 10m20.698s) test/e2e/network/loadbalancer.go:1266 In [It] (Node Runtime: 10m20.046s) test/e2e/network/loadbalancer.go:1266 At [By Step] waiting for loadbalancer for service esipp-2671/external-local-lb (Step Runtime: 10m19.729s) test/e2e/framework/service/jig.go:260 Spec Goroutine goroutine 2087 [select] k8s.io/kubernetes/vendor/k8s.io/apimachinery/pkg/util/wait.WaitForWithContext({0x7fe0bc8, 0xc00013a000}, 0xc001dd11a0, 0x2fdb16a?) vendor/k8s.io/apimachinery/pkg/util/wait/wait.go:660 k8s.io/kubernetes/vendor/k8s.io/apimachinery/pkg/util/wait.poll({0x7fe0bc8, 0xc00013a000}, 0x28?, 0x2fd9d05?, 0x20?) vendor/k8s.io/apimachinery/pkg/util/wait/wait.go:596 k8s.io/kubernetes/vendor/k8s.io/apimachinery/pkg/util/wait.PollImmediateWithContext({0x7fe0bc8, 0xc00013a000}, 0xc001534f00?, 0xc0055adb78?, 0x262a967?) vendor/k8s.io/apimachinery/pkg/util/wait/wait.go:528 k8s.io/kubernetes/vendor/k8s.io/apimachinery/pkg/util/wait.PollImmediate(0xc0010da770?, 0x7fa7740?, 0xc000214bc0?) vendor/k8s.io/apimachinery/pkg/util/wait/wait.go:514 > k8s.io/kubernetes/test/e2e/framework/service.(*TestJig).waitForCondition(0xc0047adef0, 0x4?, {0x7600fe2, 0x14}, 0x7895b68) test/e2e/framework/service/jig.go:631 > k8s.io/kubernetes/test/e2e/framework/service.(*TestJig).WaitForLoadBalancer(0xc0047adef0, 0x41?) test/e2e/framework/service/jig.go:582 > k8s.io/kubernetes/test/e2e/framework/service.(*TestJig).CreateLoadBalancerService(0xc0047adef0, 0x6aba880?, 0xc0055ade28) test/e2e/framework/service/jig.go:261 > k8s.io/kubernetes/test/e2e/framework/service.(*TestJig).CreateOnlyLocalLoadBalancerService(0xc0047adef0, 0xc002aabd40?, 0x1, 0xa?) test/e2e/framework/service/jig.go:222 > k8s.io/kubernetes/test/e2e/network.glob..func20.3() test/e2e/network/loadbalancer.go:1271 k8s.io/kubernetes/vendor/github.com/onsi/ginkgo/v2/internal.extractBodyFunction.func3({0x2d591ce, 0xc0004cfe00}) vendor/github.com/onsi/ginkgo/v2/internal/node.go:449 k8s.io/kubernetes/vendor/github.com/onsi/ginkgo/v2/internal.(*Suite).runNode.func2() vendor/github.com/onsi/ginkgo/v2/internal/suite.go:750 k8s.io/kubernetes/vendor/github.com/onsi/ginkgo/v2/internal.(*Suite).runNode vendor/github.com/onsi/ginkgo/v2/internal/suite.go:738 ------------------------------ ------------------------------ Progress Report for Ginkgo Process #11 Automatically polling progress: [sig-network] LoadBalancers ESIPP [Slow] should work for type=LoadBalancer (Spec Runtime: 10m40.737s) test/e2e/network/loadbalancer.go:1266 In [It] (Node Runtime: 10m40.085s) test/e2e/network/loadbalancer.go:1266 At [By Step] waiting for loadbalancer for service esipp-2671/external-local-lb (Step Runtime: 10m39.767s) test/e2e/framework/service/jig.go:260 Spec Goroutine goroutine 2087 [select] k8s.io/kubernetes/vendor/k8s.io/apimachinery/pkg/util/wait.WaitForWithContext({0x7fe0bc8, 0xc00013a000}, 0xc001dd11a0, 0x2fdb16a?) vendor/k8s.io/apimachinery/pkg/util/wait/wait.go:660 k8s.io/kubernetes/vendor/k8s.io/apimachinery/pkg/util/wait.poll({0x7fe0bc8, 0xc00013a000}, 0x28?, 0x2fd9d05?, 0x20?) vendor/k8s.io/apimachinery/pkg/util/wait/wait.go:596 k8s.io/kubernetes/vendor/k8s.io/apimachinery/pkg/util/wait.PollImmediateWithContext({0x7fe0bc8, 0xc00013a000}, 0xc001534f00?, 0xc0055adb78?, 0x262a967?) vendor/k8s.io/apimachinery/pkg/util/wait/wait.go:528 k8s.io/kubernetes/vendor/k8s.io/apimachinery/pkg/util/wait.PollImmediate(0xc0010da770?, 0x7fa7740?, 0xc000214bc0?) vendor/k8s.io/apimachinery/pkg/util/wait/wait.go:514 > k8s.io/kubernetes/test/e2e/framework/service.(*TestJig).waitForCondition(0xc0047adef0, 0x4?, {0x7600fe2, 0x14}, 0x7895b68) test/e2e/framework/service/jig.go:631 > k8s.io/kubernetes/test/e2e/framework/service.(*TestJig).WaitForLoadBalancer(0xc0047adef0, 0x41?) test/e2e/framework/service/jig.go:582 > k8s.io/kubernetes/test/e2e/framework/service.(*TestJig).CreateLoadBalancerService(0xc0047adef0, 0x6aba880?, 0xc0055ade28) test/e2e/framework/service/jig.go:261 > k8s.io/kubernetes/test/e2e/framework/service.(*TestJig).CreateOnlyLocalLoadBalancerService(0xc0047adef0, 0xc002aabd40?, 0x1, 0xa?) test/e2e/framework/service/jig.go:222 > k8s.io/kubernetes/test/e2e/network.glob..func20.3() test/e2e/network/loadbalancer.go:1271 k8s.io/kubernetes/vendor/github.com/onsi/ginkgo/v2/internal.extractBodyFunction.func3({0x2d591ce, 0xc0004cfe00}) vendor/github.com/onsi/ginkgo/v2/internal/node.go:449 k8s.io/kubernetes/vendor/github.com/onsi/ginkgo/v2/internal.(*Suite).runNode.func2() vendor/github.com/onsi/ginkgo/v2/internal/suite.go:750 k8s.io/kubernetes/vendor/github.com/onsi/ginkgo/v2/internal.(*Suite).runNode vendor/github.com/onsi/ginkgo/v2/internal/suite.go:738 ------------------------------ ------------------------------ Progress Report for Ginkgo Process #11 Automatically polling progress: [sig-network] LoadBalancers ESIPP [Slow] should work for type=LoadBalancer (Spec Runtime: 11m0.739s) test/e2e/network/loadbalancer.go:1266 In [It] (Node Runtime: 11m0.087s) test/e2e/network/loadbalancer.go:1266 At [By Step] waiting for loadbalancer for service esipp-2671/external-local-lb (Step Runtime: 10m59.77s) test/e2e/framework/service/jig.go:260 Spec Goroutine goroutine 2087 [select] k8s.io/kubernetes/vendor/k8s.io/apimachinery/pkg/util/wait.WaitForWithContext({0x7fe0bc8, 0xc00013a000}, 0xc001dd11a0, 0x2fdb16a?) vendor/k8s.io/apimachinery/pkg/util/wait/wait.go:660 k8s.io/kubernetes/vendor/k8s.io/apimachinery/pkg/util/wait.poll({0x7fe0bc8, 0xc00013a000}, 0x28?, 0x2fd9d05?, 0x20?) vendor/k8s.io/apimachinery/pkg/util/wait/wait.go:596 k8s.io/kubernetes/vendor/k8s.io/apimachinery/pkg/util/wait.PollImmediateWithContext({0x7fe0bc8, 0xc00013a000}, 0xc001534f00?, 0xc0055adb78?, 0x262a967?) vendor/k8s.io/apimachinery/pkg/util/wait/wait.go:528 k8s.io/kubernetes/vendor/k8s.io/apimachinery/pkg/util/wait.PollImmediate(0xc0010da770?, 0x7fa7740?, 0xc000214bc0?) vendor/k8s.io/apimachinery/pkg/util/wait/wait.go:514 > k8s.io/kubernetes/test/e2e/framework/service.(*TestJig).waitForCondition(0xc0047adef0, 0x4?, {0x7600fe2, 0x14}, 0x7895b68) test/e2e/framework/service/jig.go:631 > k8s.io/kubernetes/test/e2e/framework/service.(*TestJig).WaitForLoadBalancer(0xc0047adef0, 0x41?) test/e2e/framework/service/jig.go:582 > k8s.io/kubernetes/test/e2e/framework/service.(*TestJig).CreateLoadBalancerService(0xc0047adef0, 0x6aba880?, 0xc0055ade28) test/e2e/framework/service/jig.go:261 > k8s.io/kubernetes/test/e2e/framework/service.(*TestJig).CreateOnlyLocalLoadBalancerService(0xc0047adef0, 0xc002aabd40?, 0x1, 0xa?) test/e2e/framework/service/jig.go:222 > k8s.io/kubernetes/test/e2e/network.glob..func20.3() test/e2e/network/loadbalancer.go:1271 k8s.io/kubernetes/vendor/github.com/onsi/ginkgo/v2/internal.extractBodyFunction.func3({0x2d591ce, 0xc0004cfe00}) vendor/github.com/onsi/ginkgo/v2/internal/node.go:449 k8s.io/kubernetes/vendor/github.com/onsi/ginkgo/v2/internal.(*Suite).runNode.func2() vendor/github.com/onsi/ginkgo/v2/internal/suite.go:750 k8s.io/kubernetes/vendor/github.com/onsi/ginkgo/v2/internal.(*Suite).runNode vendor/github.com/onsi/ginkgo/v2/internal/suite.go:738 ------------------------------ ------------------------------ Progress Report for Ginkgo Process #11 Automatically polling progress: [sig-network] LoadBalancers ESIPP [Slow] should work for type=LoadBalancer (Spec Runtime: 11m20.741s) test/e2e/network/loadbalancer.go:1266 In [It] (Node Runtime: 11m20.09s) test/e2e/network/loadbalancer.go:1266 At [By Step] waiting for loadbalancer for service esipp-2671/external-local-lb (Step Runtime: 11m19.772s) test/e2e/framework/service/jig.go:260 Spec Goroutine goroutine 2087 [select] k8s.io/kubernetes/vendor/k8s.io/apimachinery/pkg/util/wait.WaitForWithContext({0x7fe0bc8, 0xc00013a000}, 0xc001dd11a0, 0x2fdb16a?) vendor/k8s.io/apimachinery/pkg/util/wait/wait.go:660 k8s.io/kubernetes/vendor/k8s.io/apimachinery/pkg/util/wait.poll({0x7fe0bc8, 0xc00013a000}, 0x28?, 0x2fd9d05?, 0x20?) vendor/k8s.io/apimachinery/pkg/util/wait/wait.go:596 k8s.io/kubernetes/vendor/k8s.io/apimachinery/pkg/util/wait.PollImmediateWithContext({0x7fe0bc8, 0xc00013a000}, 0xc001534f00?, 0xc0055adb78?, 0x262a967?) vendor/k8s.io/apimachinery/pkg/util/wait/wait.go:528 k8s.io/kubernetes/vendor/k8s.io/apimachinery/pkg/util/wait.PollImmediate(0xc0010da770?, 0x7fa7740?, 0xc000214bc0?) vendor/k8s.io/apimachinery/pkg/util/wait/wait.go:514 > k8s.io/kubernetes/test/e2e/framework/service.(*TestJig).waitForCondition(0xc0047adef0, 0x4?, {0x7600fe2, 0x14}, 0x7895b68) test/e2e/framework/service/jig.go:631 > k8s.io/kubernetes/test/e2e/framework/service.(*TestJig).WaitForLoadBalancer(0xc0047adef0, 0x41?) test/e2e/framework/service/jig.go:582 > k8s.io/kubernetes/test/e2e/framework/service.(*TestJig).CreateLoadBalancerService(0xc0047adef0, 0x6aba880?, 0xc0055ade28) test/e2e/framework/service/jig.go:261 > k8s.io/kubernetes/test/e2e/framework/service.(*TestJig).CreateOnlyLocalLoadBalancerService(0xc0047adef0, 0xc002aabd40?, 0x1, 0xa?) test/e2e/framework/service/jig.go:222 > k8s.io/kubernetes/test/e2e/network.glob..func20.3() test/e2e/network/loadbalancer.go:1271 k8s.io/kubernetes/vendor/github.com/onsi/ginkgo/v2/internal.extractBodyFunction.func3({0x2d591ce, 0xc0004cfe00}) vendor/github.com/onsi/ginkgo/v2/internal/node.go:449 k8s.io/kubernetes/vendor/github.com/onsi/ginkgo/v2/internal.(*Suite).runNode.func2() vendor/github.com/onsi/ginkgo/v2/internal/suite.go:750 k8s.io/kubernetes/vendor/github.com/onsi/ginkgo/v2/internal.(*Suite).runNode vendor/github.com/onsi/ginkgo/v2/internal/suite.go:738 ------------------------------ ------------------------------ Progress Report for Ginkgo Process #11 Automatically polling progress: [sig-network] LoadBalancers ESIPP [Slow] should work for type=LoadBalancer (Spec Runtime: 11m40.743s) test/e2e/network/loadbalancer.go:1266 In [It] (Node Runtime: 11m40.092s) test/e2e/network/loadbalancer.go:1266 At [By Step] waiting for loadbalancer for service esipp-2671/external-local-lb (Step Runtime: 11m39.774s) test/e2e/framework/service/jig.go:260 Spec Goroutine goroutine 2087 [select] k8s.io/kubernetes/vendor/k8s.io/apimachinery/pkg/util/wait.WaitForWithContext({0x7fe0bc8, 0xc00013a000}, 0xc001dd11a0, 0x2fdb16a?) vendor/k8s.io/apimachinery/pkg/util/wait/wait.go:660 k8s.io/kubernetes/vendor/k8s.io/apimachinery/pkg/util/wait.poll({0x7fe0bc8, 0xc00013a000}, 0x28?, 0x2fd9d05?, 0x20?) vendor/k8s.io/apimachinery/pkg/util/wait/wait.go:596 k8s.io/kubernetes/vendor/k8s.io/apimachinery/pkg/util/wait.PollImmediateWithContext({0x7fe0bc8, 0xc00013a000}, 0xc001534f00?, 0xc0055adb78?, 0x262a967?) vendor/k8s.io/apimachinery/pkg/util/wait/wait.go:528 k8s.io/kubernetes/vendor/k8s.io/apimachinery/pkg/util/wait.PollImmediate(0xc0010da770?, 0x7fa7740?, 0xc000214bc0?) vendor/k8s.io/apimachinery/pkg/util/wait/wait.go:514 > k8s.io/kubernetes/test/e2e/framework/service.(*TestJig).waitForCondition(0xc0047adef0, 0x4?, {0x7600fe2, 0x14}, 0x7895b68) test/e2e/framework/service/jig.go:631 > k8s.io/kubernetes/test/e2e/framework/service.(*TestJig).WaitForLoadBalancer(0xc0047adef0, 0x41?) test/e2e/framework/service/jig.go:582 > k8s.io/kubernetes/test/e2e/framework/service.(*TestJig).CreateLoadBalancerService(0xc0047adef0, 0x6aba880?, 0xc0055ade28) test/e2e/framework/service/jig.go:261 > k8s.io/kubernetes/test/e2e/framework/service.(*TestJig).CreateOnlyLocalLoadBalancerService(0xc0047adef0, 0xc002aabd40?, 0x1, 0xa?) test/e2e/framework/service/jig.go:222 > k8s.io/kubernetes/test/e2e/network.glob..func20.3() test/e2e/network/loadbalancer.go:1271 k8s.io/kubernetes/vendor/github.com/onsi/ginkgo/v2/internal.extractBodyFunction.func3({0x2d591ce, 0xc0004cfe00}) vendor/github.com/onsi/ginkgo/v2/internal/node.go:449 k8s.io/kubernetes/vendor/github.com/onsi/ginkgo/v2/internal.(*Suite).runNode.func2() vendor/github.com/onsi/ginkgo/v2/internal/suite.go:750 k8s.io/kubernetes/vendor/github.com/onsi/ginkgo/v2/internal.(*Suite).runNode vendor/github.com/onsi/ginkgo/v2/internal/suite.go:738 ------------------------------ ------------------------------ Progress Report for Ginkgo Process #11 Automatically polling progress: [sig-network] LoadBalancers ESIPP [Slow] should work for type=LoadBalancer (Spec Runtime: 12m0.746s) test/e2e/network/loadbalancer.go:1266 In [It] (Node Runtime: 12m0.094s) test/e2e/network/loadbalancer.go:1266 At [By Step] waiting for loadbalancer for service esipp-2671/external-local-lb (Step Runtime: 11m59.776s) test/e2e/framework/service/jig.go:260 Spec Goroutine goroutine 2087 [select] k8s.io/kubernetes/vendor/k8s.io/apimachinery/pkg/util/wait.WaitForWithContext({0x7fe0bc8, 0xc00013a000}, 0xc001dd11a0, 0x2fdb16a?) vendor/k8s.io/apimachinery/pkg/util/wait/wait.go:660 k8s.io/kubernetes/vendor/k8s.io/apimachinery/pkg/util/wait.poll({0x7fe0bc8, 0xc00013a000}, 0x28?, 0x2fd9d05?, 0x20?) vendor/k8s.io/apimachinery/pkg/util/wait/wait.go:596 k8s.io/kubernetes/vendor/k8s.io/apimachinery/pkg/util/wait.PollImmediateWithContext({0x7fe0bc8, 0xc00013a000}, 0xc001534f00?, 0xc0055adb78?, 0x262a967?) vendor/k8s.io/apimachinery/pkg/util/wait/wait.go:528 k8s.io/kubernetes/vendor/k8s.io/apimachinery/pkg/util/wait.PollImmediate(0xc0010da770?, 0x7fa7740?, 0xc000214bc0?) vendor/k8s.io/apimachinery/pkg/util/wait/wait.go:514 > k8s.io/kubernetes/test/e2e/framework/service.(*TestJig).waitForCondition(0xc0047adef0, 0x4?, {0x7600fe2, 0x14}, 0x7895b68) test/e2e/framework/service/jig.go:631 > k8s.io/kubernetes/test/e2e/framework/service.(*TestJig).WaitForLoadBalancer(0xc0047adef0, 0x41?) test/e2e/framework/service/jig.go:582 > k8s.io/kubernetes/test/e2e/framework/service.(*TestJig).CreateLoadBalancerService(0xc0047adef0, 0x6aba880?, 0xc0055ade28) test/e2e/framework/service/jig.go:261 > k8s.io/kubernetes/test/e2e/framework/service.(*TestJig).CreateOnlyLocalLoadBalancerService(0xc0047adef0, 0xc002aabd40?, 0x1, 0xa?) test/e2e/framework/service/jig.go:222 > k8s.io/kubernetes/test/e2e/network.glob..func20.3() test/e2e/network/loadbalancer.go:1271 k8s.io/kubernetes/vendor/github.com/onsi/ginkgo/v2/internal.extractBodyFunction.func3({0x2d591ce, 0xc0004cfe00}) vendor/github.com/onsi/ginkgo/v2/internal/node.go:449 k8s.io/kubernetes/vendor/github.com/onsi/ginkgo/v2/internal.(*Suite).runNode.func2() vendor/github.com/onsi/ginkgo/v2/internal/suite.go:750 k8s.io/kubernetes/vendor/github.com/onsi/ginkgo/v2/internal.(*Suite).runNode vendor/github.com/onsi/ginkgo/v2/internal/suite.go:738 ------------------------------ ------------------------------ Progress Report for Ginkgo Process #11 Automatically polling progress: [sig-network] LoadBalancers ESIPP [Slow] should work for type=LoadBalancer (Spec Runtime: 12m20.749s) test/e2e/network/loadbalancer.go:1266 In [It] (Node Runtime: 12m20.097s) test/e2e/network/loadbalancer.go:1266 At [By Step] waiting for loadbalancer for service esipp-2671/external-local-lb (Step Runtime: 12m19.779s) test/e2e/framework/service/jig.go:260 Spec Goroutine goroutine 2087 [select] k8s.io/kubernetes/vendor/k8s.io/apimachinery/pkg/util/wait.WaitForWithContext({0x7fe0bc8, 0xc00013a000}, 0xc001dd11a0, 0x2fdb16a?) vendor/k8s.io/apimachinery/pkg/util/wait/wait.go:660 k8s.io/kubernetes/vendor/k8s.io/apimachinery/pkg/util/wait.poll({0x7fe0bc8, 0xc00013a000}, 0x28?, 0x2fd9d05?, 0x20?) vendor/k8s.io/apimachinery/pkg/util/wait/wait.go:596 k8s.io/kubernetes/vendor/k8s.io/apimachinery/pkg/util/wait.PollImmediateWithContext({0x7fe0bc8, 0xc00013a000}, 0xc001534f00?, 0xc0055adb78?, 0x262a967?) vendor/k8s.io/apimachinery/pkg/util/wait/wait.go:528 k8s.io/kubernetes/vendor/k8s.io/apimachinery/pkg/util/wait.PollImmediate(0xc0010da770?, 0x7fa7740?, 0xc000214bc0?) vendor/k8s.io/apimachinery/pkg/util/wait/wait.go:514 > k8s.io/kubernetes/test/e2e/framework/service.(*TestJig).waitForCondition(0xc0047adef0, 0x4?, {0x7600fe2, 0x14}, 0x7895b68) test/e2e/framework/service/jig.go:631 > k8s.io/kubernetes/test/e2e/framework/service.(*TestJig).WaitForLoadBalancer(0xc0047adef0, 0x41?) test/e2e/framework/service/jig.go:582 > k8s.io/kubernetes/test/e2e/framework/service.(*TestJig).CreateLoadBalancerService(0xc0047adef0, 0x6aba880?, 0xc0055ade28) test/e2e/framework/service/jig.go:261 > k8s.io/kubernetes/test/e2e/framework/service.(*TestJig).CreateOnlyLocalLoadBalancerService(0xc0047adef0, 0xc002aabd40?, 0x1, 0xa?) test/e2e/framework/service/jig.go:222 > k8s.io/kubernetes/test/e2e/network.glob..func20.3() test/e2e/network/loadbalancer.go:1271 k8s.io/kubernetes/vendor/github.com/onsi/ginkgo/v2/internal.extractBodyFunction.func3({0x2d591ce, 0xc0004cfe00}) vendor/github.com/onsi/ginkgo/v2/internal/node.go:449 k8s.io/kubernetes/vendor/github.com/onsi/ginkgo/v2/internal.(*Suite).runNode.func2() vendor/github.com/onsi/ginkgo/v2/internal/suite.go:750 k8s.io/kubernetes/vendor/github.com/onsi/ginkgo/v2/internal.(*Suite).runNode vendor/github.com/onsi/ginkgo/v2/internal/suite.go:738 ------------------------------ ------------------------------ Progress Report for Ginkgo Process #11 Automatically polling progress: [sig-network] LoadBalancers ESIPP [Slow] should work for type=LoadBalancer (Spec Runtime: 12m40.75s) test/e2e/network/loadbalancer.go:1266 In [It] (Node Runtime: 12m40.098s) test/e2e/network/loadbalancer.go:1266 At [By Step] waiting for loadbalancer for service esipp-2671/external-local-lb (Step Runtime: 12m39.781s) test/e2e/framework/service/jig.go:260 Spec Goroutine goroutine 2087 [select] k8s.io/kubernetes/vendor/k8s.io/apimachinery/pkg/util/wait.WaitForWithContext({0x7fe0bc8, 0xc00013a000}, 0xc001dd11a0, 0x2fdb16a?) vendor/k8s.io/apimachinery/pkg/util/wait/wait.go:660 k8s.io/kubernetes/vendor/k8s.io/apimachinery/pkg/util/wait.poll({0x7fe0bc8, 0xc00013a000}, 0x28?, 0x2fd9d05?, 0x20?) vendor/k8s.io/apimachinery/pkg/util/wait/wait.go:596 k8s.io/kubernetes/vendor/k8s.io/apimachinery/pkg/util/wait.PollImmediateWithContext({0x7fe0bc8, 0xc00013a000}, 0xc001534f00?, 0xc0055adb78?, 0x262a967?) vendor/k8s.io/apimachinery/pkg/util/wait/wait.go:528 k8s.io/kubernetes/vendor/k8s.io/apimachinery/pkg/util/wait.PollImmediate(0xc0010da770?, 0x7fa7740?, 0xc000214bc0?) vendor/k8s.io/apimachinery/pkg/util/wait/wait.go:514 > k8s.io/kubernetes/test/e2e/framework/service.(*TestJig).waitForCondition(0xc0047adef0, 0x4?, {0x7600fe2, 0x14}, 0x7895b68) test/e2e/framework/service/jig.go:631 > k8s.io/kubernetes/test/e2e/framework/service.(*TestJig).WaitForLoadBalancer(0xc0047adef0, 0x41?) test/e2e/framework/service/jig.go:582 > k8s.io/kubernetes/test/e2e/framework/service.(*TestJig).CreateLoadBalancerService(0xc0047adef0, 0x6aba880?, 0xc0055ade28) test/e2e/framework/service/jig.go:261 > k8s.io/kubernetes/test/e2e/framework/service.(*TestJig).CreateOnlyLocalLoadBalancerService(0xc0047adef0, 0xc002aabd40?, 0x1, 0xa?) test/e2e/framework/service/jig.go:222 > k8s.io/kubernetes/test/e2e/network.glob..func20.3() test/e2e/network/loadbalancer.go:1271 k8s.io/kubernetes/vendor/github.com/onsi/ginkgo/v2/internal.extractBodyFunction.func3({0x2d591ce, 0xc0004cfe00}) vendor/github.com/onsi/ginkgo/v2/internal/node.go:449 k8s.io/kubernetes/vendor/github.com/onsi/ginkgo/v2/internal.(*Suite).runNode.func2() vendor/github.com/onsi/ginkgo/v2/internal/suite.go:750 k8s.io/kubernetes/vendor/github.com/onsi/ginkgo/v2/internal.(*Suite).runNode vendor/github.com/onsi/ginkgo/v2/internal/suite.go:738 ------------------------------ ------------------------------ Progress Report for Ginkgo Process #11 Automatically polling progress: [sig-network] LoadBalancers ESIPP [Slow] should work for type=LoadBalancer (Spec Runtime: 13m0.753s) test/e2e/network/loadbalancer.go:1266 In [It] (Node Runtime: 13m0.101s) test/e2e/network/loadbalancer.go:1266 At [By Step] waiting for loadbalancer for service esipp-2671/external-local-lb (Step Runtime: 12m59.783s) test/e2e/framework/service/jig.go:260 Spec Goroutine goroutine 2087 [select] k8s.io/kubernetes/vendor/k8s.io/apimachinery/pkg/util/wait.WaitForWithContext({0x7fe0bc8, 0xc00013a000}, 0xc001dd11a0, 0x2fdb16a?) vendor/k8s.io/apimachinery/pkg/util/wait/wait.go:660 k8s.io/kubernetes/vendor/k8s.io/apimachinery/pkg/util/wait.poll({0x7fe0bc8, 0xc00013a000}, 0x28?, 0x2fd9d05?, 0x20?) vendor/k8s.io/apimachinery/pkg/util/wait/wait.go:596 k8s.io/kubernetes/vendor/k8s.io/apimachinery/pkg/util/wait.PollImmediateWithContext({0x7fe0bc8, 0xc00013a000}, 0xc001534f00?, 0xc0055adb78?, 0x262a967?) vendor/k8s.io/apimachinery/pkg/util/wait/wait.go:528 k8s.io/kubernetes/vendor/k8s.io/apimachinery/pkg/util/wait.PollImmediate(0xc0010da770?, 0x7fa7740?, 0xc000214bc0?) vendor/k8s.io/apimachinery/pkg/util/wait/wait.go:514 > k8s.io/kubernetes/test/e2e/framework/service.(*TestJig).waitForCondition(0xc0047adef0, 0x4?, {0x7600fe2, 0x14}, 0x7895b68) test/e2e/framework/service/jig.go:631 > k8s.io/kubernetes/test/e2e/framework/service.(*TestJig).WaitForLoadBalancer(0xc0047adef0, 0x41?) test/e2e/framework/service/jig.go:582 > k8s.io/kubernetes/test/e2e/framework/service.(*TestJig).CreateLoadBalancerService(0xc0047adef0, 0x6aba880?, 0xc0055ade28) test/e2e/framework/service/jig.go:261 > k8s.io/kubernetes/test/e2e/framework/service.(*TestJig).CreateOnlyLocalLoadBalancerService(0xc0047adef0, 0xc002aabd40?, 0x1, 0xa?) test/e2e/framework/service/jig.go:222 > k8s.io/kubernetes/test/e2e/network.glob..func20.3() test/e2e/network/loadbalancer.go:1271 k8s.io/kubernetes/vendor/github.com/onsi/ginkgo/v2/internal.extractBodyFunction.func3({0x2d591ce, 0xc0004cfe00}) vendor/github.com/onsi/ginkgo/v2/internal/node.go:449 k8s.io/kubernetes/vendor/github.com/onsi/ginkgo/v2/internal.(*Suite).runNode.func2() vendor/github.com/onsi/ginkgo/v2/internal/suite.go:750 k8s.io/kubernetes/vendor/github.com/onsi/ginkgo/v2/internal.(*Suite).runNode vendor/github.com/onsi/ginkgo/v2/internal/suite.go:738 ------------------------------ ------------------------------ Progress Report for Ginkgo Process #11 Automatically polling progress: [sig-network] LoadBalancers ESIPP [Slow] should work for type=LoadBalancer (Spec Runtime: 13m20.76s) test/e2e/network/loadbalancer.go:1266 In [It] (Node Runtime: 13m20.108s) test/e2e/network/loadbalancer.go:1266 At [By Step] waiting for loadbalancer for service esipp-2671/external-local-lb (Step Runtime: 13m19.79s) test/e2e/framework/service/jig.go:260 Spec Goroutine goroutine 2087 [select] k8s.io/kubernetes/vendor/k8s.io/apimachinery/pkg/util/wait.WaitForWithContext({0x7fe0bc8, 0xc00013a000}, 0xc001dd11a0, 0x2fdb16a?) vendor/k8s.io/apimachinery/pkg/util/wait/wait.go:660 k8s.io/kubernetes/vendor/k8s.io/apimachinery/pkg/util/wait.poll({0x7fe0bc8, 0xc00013a000}, 0x28?, 0x2fd9d05?, 0x20?) vendor/k8s.io/apimachinery/pkg/util/wait/wait.go:596 k8s.io/kubernetes/vendor/k8s.io/apimachinery/pkg/util/wait.PollImmediateWithContext({0x7fe0bc8, 0xc00013a000}, 0xc001534f00?, 0xc0055adb78?, 0x262a967?) vendor/k8s.io/apimachinery/pkg/util/wait/wait.go:528 k8s.io/kubernetes/vendor/k8s.io/apimachinery/pkg/util/wait.PollImmediate(0xc0010da770?, 0x7fa7740?, 0xc000214bc0?) vendor/k8s.io/apimachinery/pkg/util/wait/wait.go:514 > k8s.io/kubernetes/test/e2e/framework/service.(*TestJig).waitForCondition(0xc0047adef0, 0x4?, {0x7600fe2, 0x14}, 0x7895b68) test/e2e/framework/service/jig.go:631 > k8s.io/kubernetes/test/e2e/framework/service.(*TestJig).WaitForLoadBalancer(0xc0047adef0, 0x41?) test/e2e/framework/service/jig.go:582 > k8s.io/kubernetes/test/e2e/framework/service.(*TestJig).CreateLoadBalancerService(0xc0047adef0, 0x6aba880?, 0xc0055ade28) test/e2e/framework/service/jig.go:261 > k8s.io/kubernetes/test/e2e/framework/service.(*TestJig).CreateOnlyLocalLoadBalancerService(0xc0047adef0, 0xc002aabd40?, 0x1, 0xa?) test/e2e/framework/service/jig.go:222 > k8s.io/kubernetes/test/e2e/network.glob..func20.3() test/e2e/network/loadbalancer.go:1271 k8s.io/kubernetes/vendor/github.com/onsi/ginkgo/v2/internal.extractBodyFunction.func3({0x2d591ce, 0xc0004cfe00}) vendor/github.com/onsi/ginkgo/v2/internal/node.go:449 k8s.io/kubernetes/vendor/github.com/onsi/ginkgo/v2/internal.(*Suite).runNode.func2() vendor/github.com/onsi/ginkgo/v2/internal/suite.go:750 k8s.io/kubernetes/vendor/github.com/onsi/ginkgo/v2/internal.(*Suite).runNode vendor/github.com/onsi/ginkgo/v2/internal/suite.go:738 ------------------------------ ------------------------------ Progress Report for Ginkgo Process #11 Automatically polling progress: [sig-network] LoadBalancers ESIPP [Slow] should work for type=LoadBalancer (Spec Runtime: 13m40.764s) test/e2e/network/loadbalancer.go:1266 In [It] (Node Runtime: 13m40.112s) test/e2e/network/loadbalancer.go:1266 At [By Step] waiting for loadbalancer for service esipp-2671/external-local-lb (Step Runtime: 13m39.795s) test/e2e/framework/service/jig.go:260 Spec Goroutine goroutine 2087 [select] k8s.io/kubernetes/vendor/k8s.io/apimachinery/pkg/util/wait.WaitForWithContext({0x7fe0bc8, 0xc00013a000}, 0xc001dd11a0, 0x2fdb16a?) vendor/k8s.io/apimachinery/pkg/util/wait/wait.go:660 k8s.io/kubernetes/vendor/k8s.io/apimachinery/pkg/util/wait.poll({0x7fe0bc8, 0xc00013a000}, 0x28?, 0x2fd9d05?, 0x20?) vendor/k8s.io/apimachinery/pkg/util/wait/wait.go:596 k8s.io/kubernetes/vendor/k8s.io/apimachinery/pkg/util/wait.PollImmediateWithContext({0x7fe0bc8, 0xc00013a000}, 0xc001534f00?, 0xc0055adb78?, 0x262a967?) vendor/k8s.io/apimachinery/pkg/util/wait/wait.go:528 k8s.io/kubernetes/vendor/k8s.io/apimachinery/pkg/util/wait.PollImmediate(0xc0010da770?, 0x7fa7740?, 0xc000214bc0?) vendor/k8s.io/apimachinery/pkg/util/wait/wait.go:514 > k8s.io/kubernetes/test/e2e/framework/service.(*TestJig).waitForCondition(0xc0047adef0, 0x4?, {0x7600fe2, 0x14}, 0x7895b68) test/e2e/framework/service/jig.go:631 > k8s.io/kubernetes/test/e2e/framework/service.(*TestJig).WaitForLoadBalancer(0xc0047adef0, 0x41?) test/e2e/framework/service/jig.go:582 > k8s.io/kubernetes/test/e2e/framework/service.(*TestJig).CreateLoadBalancerService(0xc0047adef0, 0x6aba880?, 0xc0055ade28) test/e2e/framework/service/jig.go:261 > k8s.io/kubernetes/test/e2e/framework/service.(*TestJig).CreateOnlyLocalLoadBalancerService(0xc0047adef0, 0xc002aabd40?, 0x1, 0xa?) test/e2e/framework/service/jig.go:222 > k8s.io/kubernetes/test/e2e/network.glob..func20.3() test/e2e/network/loadbalancer.go:1271 k8s.io/kubernetes/vendor/github.com/onsi/ginkgo/v2/internal.extractBodyFunction.func3({0x2d591ce, 0xc0004cfe00}) vendor/github.com/onsi/ginkgo/v2/internal/node.go:449 k8s.io/kubernetes/vendor/github.com/onsi/ginkgo/v2/internal.(*Suite).runNode.func2() vendor/github.com/onsi/ginkgo/v2/internal/suite.go:750 k8s.io/kubernetes/vendor/github.com/onsi/ginkgo/v2/internal.(*Suite).runNode vendor/github.com/onsi/ginkgo/v2/internal/suite.go:738 ------------------------------ ------------------------------ Progress Report for Ginkgo Process #11 Automatically polling progress: [sig-network] LoadBalancers ESIPP [Slow] should work for type=LoadBalancer (Spec Runtime: 14m0.767s) test/e2e/network/loadbalancer.go:1266 In [It] (Node Runtime: 14m0.115s) test/e2e/network/loadbalancer.go:1266 At [By Step] waiting for loadbalancer for service esipp-2671/external-local-lb (Step Runtime: 13m59.798s) test/e2e/framework/service/jig.go:260 Spec Goroutine goroutine 2087 [select] k8s.io/kubernetes/vendor/k8s.io/apimachinery/pkg/util/wait.WaitForWithContext({0x7fe0bc8, 0xc00013a000}, 0xc001dd11a0, 0x2fdb16a?) vendor/k8s.io/apimachinery/pkg/util/wait/wait.go:660 k8s.io/kubernetes/vendor/k8s.io/apimachinery/pkg/util/wait.poll({0x7fe0bc8, 0xc00013a000}, 0x28?, 0x2fd9d05?, 0x20?) vendor/k8s.io/apimachinery/pkg/util/wait/wait.go:596 k8s.io/kubernetes/vendor/k8s.io/apimachinery/pkg/util/wait.PollImmediateWithContext({0x7fe0bc8, 0xc00013a000}, 0xc001534f00?, 0xc0055adb78?, 0x262a967?) vendor/k8s.io/apimachinery/pkg/util/wait/wait.go:528 k8s.io/kubernetes/vendor/k8s.io/apimachinery/pkg/util/wait.PollImmediate(0xc0010da770?, 0x7fa7740?, 0xc000214bc0?) vendor/k8s.io/apimachinery/pkg/util/wait/wait.go:514 > k8s.io/kubernetes/test/e2e/framework/service.(*TestJig).waitForCondition(0xc0047adef0, 0x4?, {0x7600fe2, 0x14}, 0x7895b68) test/e2e/framework/service/jig.go:631 > k8s.io/kubernetes/test/e2e/framework/service.(*TestJig).WaitForLoadBalancer(0xc0047adef0, 0x41?) test/e2e/framework/service/jig.go:582 > k8s.io/kubernetes/test/e2e/framework/service.(*TestJig).CreateLoadBalancerService(0xc0047adef0, 0x6aba880?, 0xc0055ade28) test/e2e/framework/service/jig.go:261 > k8s.io/kubernetes/test/e2e/framework/service.(*TestJig).CreateOnlyLocalLoadBalancerService(0xc0047adef0, 0xc002aabd40?, 0x1, 0xa?) test/e2e/framework/service/jig.go:222 > k8s.io/kubernetes/test/e2e/network.glob..func20.3() test/e2e/network/loadbalancer.go:1271 k8s.io/kubernetes/vendor/github.com/onsi/ginkgo/v2/internal.extractBodyFunction.func3({0x2d591ce, 0xc0004cfe00}) vendor/github.com/onsi/ginkgo/v2/internal/node.go:449 k8s.io/kubernetes/vendor/github.com/onsi/ginkgo/v2/internal.(*Suite).runNode.func2() vendor/github.com/onsi/ginkgo/v2/internal/suite.go:750 k8s.io/kubernetes/vendor/github.com/onsi/ginkgo/v2/internal.(*Suite).runNode vendor/github.com/onsi/ginkgo/v2/internal/suite.go:738 ------------------------------ Nov 26 05:31:45.490: INFO: Retrying .... error trying to get Service external-local-lb: Get "https://35.227.179.144/api/v1/namespaces/esipp-2671/services/external-local-lb": dial tcp 35.227.179.144:443: connect: connection refused Nov 26 05:31:47.491: INFO: Retrying .... error trying to get Service external-local-lb: Get "https://35.227.179.144/api/v1/namespaces/esipp-2671/services/external-local-lb": dial tcp 35.227.179.144:443: connect: connection refused Nov 26 05:31:49.491: INFO: Retrying .... error trying to get Service external-local-lb: Get "https://35.227.179.144/api/v1/namespaces/esipp-2671/services/external-local-lb": dial tcp 35.227.179.144:443: connect: connection refused Nov 26 05:31:51.490: INFO: Retrying .... error trying to get Service external-local-lb: Get "https://35.227.179.144/api/v1/namespaces/esipp-2671/services/external-local-lb": dial tcp 35.227.179.144:443: connect: connection refused ------------------------------ Progress Report for Ginkgo Process #11 Automatically polling progress: [sig-network] LoadBalancers ESIPP [Slow] should work for type=LoadBalancer (Spec Runtime: 14m20.773s) test/e2e/network/loadbalancer.go:1266 In [It] (Node Runtime: 14m20.121s) test/e2e/network/loadbalancer.go:1266 At [By Step] waiting for loadbalancer for service esipp-2671/external-local-lb (Step Runtime: 14m19.804s) test/e2e/framework/service/jig.go:260 Spec Goroutine goroutine 2087 [select] k8s.io/kubernetes/vendor/k8s.io/apimachinery/pkg/util/wait.WaitForWithContext({0x7fe0bc8, 0xc00013a000}, 0xc001dd11a0, 0x2fdb16a?) vendor/k8s.io/apimachinery/pkg/util/wait/wait.go:660 k8s.io/kubernetes/vendor/k8s.io/apimachinery/pkg/util/wait.poll({0x7fe0bc8, 0xc00013a000}, 0x28?, 0x2fd9d05?, 0x20?) vendor/k8s.io/apimachinery/pkg/util/wait/wait.go:596 k8s.io/kubernetes/vendor/k8s.io/apimachinery/pkg/util/wait.PollImmediateWithContext({0x7fe0bc8, 0xc00013a000}, 0xc001534f00?, 0xc0055adb78?, 0x262a967?) vendor/k8s.io/apimachinery/pkg/util/wait/wait.go:528 k8s.io/kubernetes/vendor/k8s.io/apimachinery/pkg/util/wait.PollImmediate(0xc0010da770?, 0x7fa7740?, 0xc000214bc0?) vendor/k8s.io/apimachinery/pkg/util/wait/wait.go:514 > k8s.io/kubernetes/test/e2e/framework/service.(*TestJig).waitForCondition(0xc0047adef0, 0x4?, {0x7600fe2, 0x14}, 0x7895b68) test/e2e/framework/service/jig.go:631 > k8s.io/kubernetes/test/e2e/framework/service.(*TestJig).WaitForLoadBalancer(0xc0047adef0, 0x41?) test/e2e/framework/service/jig.go:582 > k8s.io/kubernetes/test/e2e/framework/service.(*TestJig).CreateLoadBalancerService(0xc0047adef0, 0x6aba880?, 0xc0055ade28) test/e2e/framework/service/jig.go:261 > k8s.io/kubernetes/test/e2e/framework/service.(*TestJig).CreateOnlyLocalLoadBalancerService(0xc0047adef0, 0xc002aabd40?, 0x1, 0xa?) test/e2e/framework/service/jig.go:222 > k8s.io/kubernetes/test/e2e/network.glob..func20.3() test/e2e/network/loadbalancer.go:1271 k8s.io/kubernetes/vendor/github.com/onsi/ginkgo/v2/internal.extractBodyFunction.func3({0x2d591ce, 0xc0004cfe00}) vendor/github.com/onsi/ginkgo/v2/internal/node.go:449 k8s.io/kubernetes/vendor/github.com/onsi/ginkgo/v2/internal.(*Suite).runNode.func2() vendor/github.com/onsi/ginkgo/v2/internal/suite.go:750 k8s.io/kubernetes/vendor/github.com/onsi/ginkgo/v2/internal.(*Suite).runNode vendor/github.com/onsi/ginkgo/v2/internal/suite.go:738 ------------------------------ Nov 26 05:31:53.490: INFO: Retrying .... error trying to get Service external-local-lb: Get "https://35.227.179.144/api/v1/namespaces/esipp-2671/services/external-local-lb": dial tcp 35.227.179.144:443: connect: connection refused Nov 26 05:31:55.491: INFO: Retrying .... error trying to get Service external-local-lb: Get "https://35.227.179.144/api/v1/namespaces/esipp-2671/services/external-local-lb": dial tcp 35.227.179.144:443: connect: connection refused Nov 26 05:31:57.490: INFO: Retrying .... error trying to get Service external-local-lb: Get "https://35.227.179.144/api/v1/namespaces/esipp-2671/services/external-local-lb": dial tcp 35.227.179.144:443: connect: connection refused Nov 26 05:31:59.491: INFO: Retrying .... error trying to get Service external-local-lb: Get "https://35.227.179.144/api/v1/namespaces/esipp-2671/services/external-local-lb": dial tcp 35.227.179.144:443: connect: connection refused Nov 26 05:32:01.491: INFO: Retrying .... error trying to get Service external-local-lb: Get "https://35.227.179.144/api/v1/namespaces/esipp-2671/services/external-local-lb": dial tcp 35.227.179.144:443: connect: connection refused Nov 26 05:32:03.491: INFO: Retrying .... error trying to get Service external-local-lb: Get "https://35.227.179.144/api/v1/namespaces/esipp-2671/services/external-local-lb": dial tcp 35.227.179.144:443: connect: connection refused Nov 26 05:32:05.491: INFO: Retrying .... error trying to get Service external-local-lb: Get "https://35.227.179.144/api/v1/namespaces/esipp-2671/services/external-local-lb": dial tcp 35.227.179.144:443: connect: connection refused Nov 26 05:32:07.491: INFO: Retrying .... error trying to get Service external-local-lb: Get "https://35.227.179.144/api/v1/namespaces/esipp-2671/services/external-local-lb": dial tcp 35.227.179.144:443: connect: connection refused Nov 26 05:32:09.490: INFO: Retrying .... error trying to get Service external-local-lb: Get "https://35.227.179.144/api/v1/namespaces/esipp-2671/services/external-local-lb": dial tcp 35.227.179.144:443: connect: connection refused Nov 26 05:32:11.490: INFO: Retrying .... error trying to get Service external-local-lb: Get "https://35.227.179.144/api/v1/namespaces/esipp-2671/services/external-local-lb": dial tcp 35.227.179.144:443: connect: connection refused ------------------------------ Progress Report for Ginkgo Process #11 Automatically polling progress: [sig-network] LoadBalancers ESIPP [Slow] should work for type=LoadBalancer (Spec Runtime: 14m40.775s) test/e2e/network/loadbalancer.go:1266 In [It] (Node Runtime: 14m40.123s) test/e2e/network/loadbalancer.go:1266 At [By Step] waiting for loadbalancer for service esipp-2671/external-local-lb (Step Runtime: 14m39.806s) test/e2e/framework/service/jig.go:260 Spec Goroutine goroutine 2087 [select] k8s.io/kubernetes/vendor/k8s.io/apimachinery/pkg/util/wait.WaitForWithContext({0x7fe0bc8, 0xc00013a000}, 0xc001dd11a0, 0x2fdb16a?) vendor/k8s.io/apimachinery/pkg/util/wait/wait.go:660 k8s.io/kubernetes/vendor/k8s.io/apimachinery/pkg/util/wait.poll({0x7fe0bc8, 0xc00013a000}, 0x28?, 0x2fd9d05?, 0x20?) vendor/k8s.io/apimachinery/pkg/util/wait/wait.go:596 k8s.io/kubernetes/vendor/k8s.io/apimachinery/pkg/util/wait.PollImmediateWithContext({0x7fe0bc8, 0xc00013a000}, 0xc001534f00?, 0xc0055adb78?, 0x262a967?) vendor/k8s.io/apimachinery/pkg/util/wait/wait.go:528 k8s.io/kubernetes/vendor/k8s.io/apimachinery/pkg/util/wait.PollImmediate(0xc0010da770?, 0x7fa7740?, 0xc000214bc0?) vendor/k8s.io/apimachinery/pkg/util/wait/wait.go:514 > k8s.io/kubernetes/test/e2e/framework/service.(*TestJig).waitForCondition(0xc0047adef0, 0x4?, {0x7600fe2, 0x14}, 0x7895b68) test/e2e/framework/service/jig.go:631 > k8s.io/kubernetes/test/e2e/framework/service.(*TestJig).WaitForLoadBalancer(0xc0047adef0, 0x41?) test/e2e/framework/service/jig.go:582 > k8s.io/kubernetes/test/e2e/framework/service.(*TestJig).CreateLoadBalancerService(0xc0047adef0, 0x6aba880?, 0xc0055ade28) test/e2e/framework/service/jig.go:261 > k8s.io/kubernetes/test/e2e/framework/service.(*TestJig).CreateOnlyLocalLoadBalancerService(0xc0047adef0, 0xc002aabd40?, 0x1, 0xa?) test/e2e/framework/service/jig.go:222 > k8s.io/kubernetes/test/e2e/network.glob..func20.3() test/e2e/network/loadbalancer.go:1271 k8s.io/kubernetes/vendor/github.com/onsi/ginkgo/v2/internal.extractBodyFunction.func3({0x2d591ce, 0xc0004cfe00}) vendor/github.com/onsi/ginkgo/v2/internal/node.go:449 k8s.io/kubernetes/vendor/github.com/onsi/ginkgo/v2/internal.(*Suite).runNode.func2() vendor/github.com/onsi/ginkgo/v2/internal/suite.go:750 k8s.io/kubernetes/vendor/github.com/onsi/ginkgo/v2/internal.(*Suite).runNode vendor/github.com/onsi/ginkgo/v2/internal/suite.go:738 ------------------------------ Nov 26 05:32:13.490: INFO: Retrying .... error trying to get Service external-local-lb: Get "https://35.227.179.144/api/v1/namespaces/esipp-2671/services/external-local-lb": dial tcp 35.227.179.144:443: connect: connection refused Nov 26 05:32:15.491: INFO: Retrying .... error trying to get Service external-local-lb: Get "https://35.227.179.144/api/v1/namespaces/esipp-2671/services/external-local-lb": dial tcp 35.227.179.144:443: connect: connection refused Nov 26 05:32:17.490: INFO: Retrying .... error trying to get Service external-local-lb: Get "https://35.227.179.144/api/v1/namespaces/esipp-2671/services/external-local-lb": dial tcp 35.227.179.144:443: connect: connection refused Nov 26 05:32:19.491: INFO: Retrying .... error trying to get Service external-local-lb: Get "https://35.227.179.144/api/v1/namespaces/esipp-2671/services/external-local-lb": dial tcp 35.227.179.144:443: connect: connection refused Nov 26 05:32:21.491: INFO: Retrying .... error trying to get Service external-local-lb: Get "https://35.227.179.144/api/v1/namespaces/esipp-2671/services/external-local-lb": dial tcp 35.227.179.144:443: connect: connection refused Nov 26 05:32:23.490: INFO: Retrying .... error trying to get Service external-local-lb: Get "https://35.227.179.144/api/v1/namespaces/esipp-2671/services/external-local-lb": dial tcp 35.227.179.144:443: connect: connection refused Nov 26 05:32:25.491: INFO: Retrying .... error trying to get Service external-local-lb: Get "https://35.227.179.144/api/v1/namespaces/esipp-2671/services/external-local-lb": dial tcp 35.227.179.144:443: connect: connection refused Nov 26 05:32:27.491: INFO: Retrying .... error trying to get Service external-local-lb: Get "https://35.227.179.144/api/v1/namespaces/esipp-2671/services/external-local-lb": dial tcp 35.227.179.144:443: connect: connection refused ------------------------------ Progress Report for Ginkgo Process #11 Automatically polling progress: [sig-network] LoadBalancers ESIPP [Slow] should work for type=LoadBalancer (Spec Runtime: 15m0.777s) test/e2e/network/loadbalancer.go:1266 In [It] (Node Runtime: 15m0.125s) test/e2e/network/loadbalancer.go:1266 At [By Step] waiting for loadbalancer for service esipp-2671/external-local-lb (Step Runtime: 14m59.807s) test/e2e/framework/service/jig.go:260 Spec Goroutine goroutine 2087 [select] k8s.io/kubernetes/vendor/k8s.io/apimachinery/pkg/util/wait.WaitForWithContext({0x7fe0bc8, 0xc00013a000}, 0xc001dd11a0, 0x2fdb16a?) vendor/k8s.io/apimachinery/pkg/util/wait/wait.go:660 k8s.io/kubernetes/vendor/k8s.io/apimachinery/pkg/util/wait.poll({0x7fe0bc8, 0xc00013a000}, 0x28?, 0x2fd9d05?, 0x20?) vendor/k8s.io/apimachinery/pkg/util/wait/wait.go:596 k8s.io/kubernetes/vendor/k8s.io/apimachinery/pkg/util/wait.PollImmediateWithContext({0x7fe0bc8, 0xc00013a000}, 0xc001534f00?, 0xc0055adb78?, 0x262a967?) vendor/k8s.io/apimachinery/pkg/util/wait/wait.go:528 k8s.io/kubernetes/vendor/k8s.io/apimachinery/pkg/util/wait.PollImmediate(0xc0010da770?, 0x7fa7740?, 0xc000214bc0?) vendor/k8s.io/apimachinery/pkg/util/wait/wait.go:514 > k8s.io/kubernetes/test/e2e/framework/service.(*TestJig).waitForCondition(0xc0047adef0, 0x4?, {0x7600fe2, 0x14}, 0x7895b68) test/e2e/framework/service/jig.go:631 > k8s.io/kubernetes/test/e2e/framework/service.(*TestJig).WaitForLoadBalancer(0xc0047adef0, 0x41?) test/e2e/framework/service/jig.go:582 > k8s.io/kubernetes/test/e2e/framework/service.(*TestJig).CreateLoadBalancerService(0xc0047adef0, 0x6aba880?, 0xc0055ade28) test/e2e/framework/service/jig.go:261 > k8s.io/kubernetes/test/e2e/framework/service.(*TestJig).CreateOnlyLocalLoadBalancerService(0xc0047adef0, 0xc002aabd40?, 0x1, 0xa?) test/e2e/framework/service/jig.go:222 > k8s.io/kubernetes/test/e2e/network.glob..func20.3() test/e2e/network/loadbalancer.go:1271 k8s.io/kubernetes/vendor/github.com/onsi/ginkgo/v2/internal.extractBodyFunction.func3({0x2d591ce, 0xc0004cfe00}) vendor/github.com/onsi/ginkgo/v2/internal/node.go:449 k8s.io/kubernetes/vendor/github.com/onsi/ginkgo/v2/internal.(*Suite).runNode.func2() vendor/github.com/onsi/ginkgo/v2/internal/suite.go:750 k8s.io/kubernetes/vendor/github.com/onsi/ginkgo/v2/internal.(*Suite).runNode vendor/github.com/onsi/ginkgo/v2/internal/suite.go:738 ------------------------------ Nov 26 05:32:33.538: INFO: Unexpected error: <*fmt.wrapError | 0xc0002191c0>: { msg: "timed out waiting for service \"external-local-lb\" to have a load balancer: timed out waiting for the condition", err: <*errors.errorString | 0xc00020fc70>{ s: "timed out waiting for the condition", }, } Nov 26 05:32:33.538: FAIL: timed out waiting for service "external-local-lb" to have a load balancer: timed out waiting for the condition Full Stack Trace k8s.io/kubernetes/test/e2e/network.glob..func20.3() test/e2e/network/loadbalancer.go:1272 +0xd8 [AfterEach] [sig-network] LoadBalancers ESIPP [Slow] test/e2e/framework/node/init/init.go:32 Nov 26 05:32:33.538: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready [AfterEach] [sig-network] LoadBalancers ESIPP [Slow] test/e2e/network/loadbalancer.go:1260 Nov 26 05:32:33.621: INFO: Output of kubectl describe svc: Nov 26 05:32:33.621: INFO: Running '/workspace/github.com/containerd/containerd/kubernetes/platforms/linux/amd64/kubectl --server=https://35.227.179.144 --kubeconfig=/workspace/.kube/config --namespace=esipp-2671 describe svc --namespace=esipp-2671' Nov 26 05:32:33.952: INFO: stderr: "" Nov 26 05:32:33.952: INFO: stdout: "Name: external-local-lb\nNamespace: esipp-2671\nLabels: testid=external-local-lb-f345919d-89e5-4713-806c-87373fa4a70a\nAnnotations: <none>\nSelector: testid=external-local-lb-f345919d-89e5-4713-806c-87373fa4a70a\nType: LoadBalancer\nIP Family Policy: SingleStack\nIP Families: IPv4\nIP: 10.0.202.60\nIPs: 10.0.202.60\nPort: <unset> 80/TCP\nTargetPort: 80/TCP\nNodePort: <unset> 31630/TCP\nEndpoints: <none>\nSession Affinity: None\nExternal Traffic Policy: Local\nHealthCheck NodePort: 32108\nEvents: <none>\n" Nov 26 05:32:33.952: INFO: Name: external-local-lb Namespace: esipp-2671 Labels: testid=external-local-lb-f345919d-89e5-4713-806c-87373fa4a70a Annotations: <none> Selector: testid=external-local-lb-f345919d-89e5-4713-806c-87373fa4a70a Type: LoadBalancer IP Family Policy: SingleStack IP Families: IPv4 IP: 10.0.202.60 IPs: 10.0.202.60 Port: <unset> 80/TCP TargetPort: 80/TCP NodePort: <unset> 31630/TCP Endpoints: <none> Session Affinity: None External Traffic Policy: Local HealthCheck NodePort: 32108 Events: <none> [DeferCleanup (Each)] [sig-network] LoadBalancers ESIPP [Slow] test/e2e/framework/metrics/init/init.go:33 [DeferCleanup (Each)] [sig-network] LoadBalancers ESIPP [Slow] dump namespaces | framework.go:196 STEP: dump namespace information after failure 11/26/22 05:32:33.952 STEP: Collecting events from namespace "esipp-2671". 11/26/22 05:32:33.952 STEP: Found 0 events. 11/26/22 05:32:33.992 Nov 26 05:32:34.033: INFO: POD NODE PHASE GRACE CONDITIONS Nov 26 05:32:34.033: INFO: Nov 26 05:32:34.082: INFO: Logging node info for node bootstrap-e2e-master Nov 26 05:32:34.123: INFO: Node Info: &Node{ObjectMeta:{bootstrap-e2e-master 9a48eb11-c964-4bf8-82c0-1c32a9703981 10713 0 2022-11-26 05:05:18 +0000 UTC <nil> <nil> map[beta.kubernetes.io/arch:amd64 beta.kubernetes.io/instance-type:n1-standard-1 beta.kubernetes.io/os:linux cloud.google.com/metadata-proxy-ready:true failure-domain.beta.kubernetes.io/region:us-west1 failure-domain.beta.kubernetes.io/zone:us-west1-b kubernetes.io/arch:amd64 kubernetes.io/hostname:bootstrap-e2e-master kubernetes.io/os:linux node.kubernetes.io/instance-type:n1-standard-1 topology.kubernetes.io/region:us-west1 topology.kubernetes.io/zone:us-west1-b] map[node.alpha.kubernetes.io/ttl:0 volumes.kubernetes.io/controller-managed-attach-detach:true] [] [] [{kubelet Update v1 2022-11-26 05:05:18 +0000 UTC FieldsV1 {"f:metadata":{"f:annotations":{".":{},"f:volumes.kubernetes.io/controller-managed-attach-detach":{}},"f:labels":{".":{},"f:beta.kubernetes.io/arch":{},"f:beta.kubernetes.io/instance-type":{},"f:beta.kubernetes.io/os":{},"f:cloud.google.com/metadata-proxy-ready":{},"f:failure-domain.beta.kubernetes.io/region":{},"f:failure-domain.beta.kubernetes.io/zone":{},"f:kubernetes.io/arch":{},"f:kubernetes.io/hostname":{},"f:kubernetes.io/os":{},"f:node.kubernetes.io/instance-type":{},"f:topology.kubernetes.io/region":{},"f:topology.kubernetes.io/zone":{}}},"f:spec":{"f:providerID":{},"f:unschedulable":{}}} } {kube-controller-manager Update v1 2022-11-26 05:05:25 +0000 UTC FieldsV1 {"f:status":{"f:conditions":{"k:{\"type\":\"NetworkUnavailable\"}":{"f:lastHeartbeatTime":{},"f:lastTransitionTime":{},"f:message":{},"f:reason":{},"f:status":{}}}}} status} {kube-controller-manager Update v1 2022-11-26 05:05:39 +0000 UTC FieldsV1 {"f:metadata":{"f:annotations":{"f:node.alpha.kubernetes.io/ttl":{}}},"f:spec":{"f:podCIDR":{},"f:podCIDRs":{".":{},"v:\"10.64.1.0/24\"":{}},"f:taints":{}}} } {kubelet Update v1 2022-11-26 05:27:29 +0000 UTC FieldsV1 {"f:status":{"f:conditions":{"k:{\"type\":\"DiskPressure\"}":{"f:lastHeartbeatTime":{}},"k:{\"type\":\"MemoryPressure\"}":{"f:lastHeartbeatTime":{}},"k:{\"type\":\"PIDPressure\"}":{"f:lastHeartbeatTime":{}},"k:{\"type\":\"Ready\"}":{"f:lastHeartbeatTime":{},"f:lastTransitionTime":{},"f:message":{},"f:reason":{},"f:status":{}}},"f:images":{}}} status}]},Spec:NodeSpec{PodCIDR:10.64.1.0/24,DoNotUseExternalID:,ProviderID:gce://k8s-jkns-e2e-gce-ci-reboot/us-west1-b/bootstrap-e2e-master,Unschedulable:true,Taints:[]Taint{Taint{Key:node-role.kubernetes.io/master,Value:,Effect:NoSchedule,TimeAdded:<nil>,},Taint{Key:node.kubernetes.io/unschedulable,Value:,Effect:NoSchedule,TimeAdded:<nil>,},},ConfigSource:nil,PodCIDRs:[10.64.1.0/24],},Status:NodeStatus{Capacity:ResourceList{attachable-volumes-gce-pd: {{127 0} {<nil>} 127 DecimalSI},cpu: {{1 0} {<nil>} 1 DecimalSI},ephemeral-storage: {{16656896000 0} {<nil>} 16266500Ki BinarySI},hugepages-1Gi: {{0 0} {<nil>} 0 DecimalSI},hugepages-2Mi: {{0 0} {<nil>} 0 DecimalSI},memory: {{3858374656 0} {<nil>} BinarySI},pods: {{110 0} {<nil>} 110 DecimalSI},},Allocatable:ResourceList{attachable-volumes-gce-pd: {{127 0} {<nil>} 127 DecimalSI},cpu: {{1 0} {<nil>} 1 DecimalSI},ephemeral-storage: {{14991206376 0} {<nil>} 14991206376 DecimalSI},hugepages-1Gi: {{0 0} {<nil>} 0 DecimalSI},hugepages-2Mi: {{0 0} {<nil>} 0 DecimalSI},memory: {{3596230656 0} {<nil>} BinarySI},pods: {{110 0} {<nil>} 110 DecimalSI},},Phase:,Conditions:[]NodeCondition{NodeCondition{Type:NetworkUnavailable,Status:False,LastHeartbeatTime:2022-11-26 05:05:25 +0000 UTC,LastTransitionTime:2022-11-26 05:05:25 +0000 UTC,Reason:RouteCreated,Message:RouteController created a route,},NodeCondition{Type:MemoryPressure,Status:False,LastHeartbeatTime:2022-11-26 05:27:29 +0000 UTC,LastTransitionTime:2022-11-26 05:05:18 +0000 UTC,Reason:KubeletHasSufficientMemory,Message:kubelet has sufficient memory available,},NodeCondition{Type:DiskPressure,Status:False,LastHeartbeatTime:2022-11-26 05:27:29 +0000 UTC,LastTransitionTime:2022-11-26 05:05:18 +0000 UTC,Reason:KubeletHasNoDiskPressure,Message:kubelet has no disk pressure,},NodeCondition{Type:PIDPressure,Status:False,LastHeartbeatTime:2022-11-26 05:27:29 +0000 UTC,LastTransitionTime:2022-11-26 05:05:18 +0000 UTC,Reason:KubeletHasSufficientPID,Message:kubelet has sufficient PID available,},NodeCondition{Type:Ready,Status:True,LastHeartbeatTime:2022-11-26 05:27:29 +0000 UTC,LastTransitionTime:2022-11-26 05:05:38 +0000 UTC,Reason:KubeletReady,Message:kubelet is posting ready status. AppArmor enabled,},},Addresses:[]NodeAddress{NodeAddress{Type:InternalIP,Address:10.138.0.2,},NodeAddress{Type:ExternalIP,Address:35.227.179.144,},NodeAddress{Type:InternalDNS,Address:bootstrap-e2e-master.c.k8s-jkns-e2e-gce-ci-reboot.internal,},NodeAddress{Type:Hostname,Address:bootstrap-e2e-master.c.k8s-jkns-e2e-gce-ci-reboot.internal,},},DaemonEndpoints:NodeDaemonEndpoints{KubeletEndpoint:DaemonEndpoint{Port:10250,},},NodeInfo:NodeSystemInfo{MachineID:0fdb3cfe29f66637553465718381a2f8,SystemUUID:0fdb3cfe-29f6-6637-5534-65718381a2f8,BootID:fb804a29-7e61-434d-8044-6944159ce35f,KernelVersion:5.10.123+,OSImage:Container-Optimized OS from Google,ContainerRuntimeVersion:containerd://1.7.0-beta.0-149-gd06318622,KubeletVersion:v1.27.0-alpha.0.50+70617042976dc1,KubeProxyVersion:v1.27.0-alpha.0.50+70617042976dc1,OperatingSystem:linux,Architecture:amd64,},Images:[]ContainerImage{ContainerImage{Names:[registry.k8s.io/kube-apiserver-amd64:v1.27.0-alpha.0.50_70617042976dc1],SizeBytes:135160272,},ContainerImage{Names:[registry.k8s.io/kube-controller-manager-amd64:v1.27.0-alpha.0.50_70617042976dc1],SizeBytes:124990265,},ContainerImage{Names:[registry.k8s.io/etcd@sha256:dd75ec974b0a2a6f6bb47001ba09207976e625db898d1b16735528c009cb171c registry.k8s.io/etcd:3.5.6-0],SizeBytes:102542580,},ContainerImage{Names:[registry.k8s.io/kube-scheduler-amd64:v1.27.0-alpha.0.50_70617042976dc1],SizeBytes:57660216,},ContainerImage{Names:[gke.gcr.io/prometheus-to-sd@sha256:e739643c3939ba0b161425f45a1989eedfc4a3b166db9a7100863296b4c70510 gke.gcr.io/prometheus-to-sd:v0.11.1-gke.1],SizeBytes:48742566,},ContainerImage{Names:[gcr.io/k8s-ingress-image-push/ingress-gce-glbc-amd64@sha256:5db27383add6d9f4ebdf0286409ac31f7f5d273690204b341a4e37998917693b gcr.io/k8s-ingress-image-push/ingress-gce-glbc-amd64:v1.20.1],SizeBytes:36598135,},ContainerImage{Names:[registry.k8s.io/addon-manager/kube-addon-manager@sha256:49cc4e6e4a3745b427ce14b0141476ab339bb65c6bc05033019e046c8727dcb0 registry.k8s.io/addon-manager/kube-addon-manager:v9.1.6],SizeBytes:30464183,},ContainerImage{Names:[registry.k8s.io/kas-network-proxy/proxy-server@sha256:2c111f004bec24888d8cfa2a812a38fb8341350abac67dcd0ac64e709dfe389c registry.k8s.io/kas-network-proxy/proxy-server:v0.0.33],SizeBytes:22020129,},ContainerImage{Names:[registry.k8s.io/metadata-proxy@sha256:e914645f22e946bce5165737e1b244e0a296ad1f0f81a9531adc57af2780978a registry.k8s.io/metadata-proxy:v0.1.12],SizeBytes:5301657,},ContainerImage{Names:[registry.k8s.io/pause@sha256:9001185023633d17a2f98ff69b6ff2615b8ea02a825adffa40422f51dfdcde9d registry.k8s.io/pause:3.8],SizeBytes:311286,},},VolumesInUse:[],VolumesAttached:[]AttachedVolume{},Config:nil,},} Nov 26 05:32:34.124: INFO: Logging kubelet events for node bootstrap-e2e-master Nov 26 05:32:34.178: INFO: Logging pods the kubelet thinks is on node bootstrap-e2e-master Nov 26 05:32:34.223: INFO: Unable to retrieve kubelet pods for node bootstrap-e2e-master: error trying to reach service: No agent available Nov 26 05:32:34.223: INFO: Logging node info for node bootstrap-e2e-minion-group-0975 Nov 26 05:32:34.269: INFO: Node Info: &Node{ObjectMeta:{bootstrap-e2e-minion-group-0975 6a0671d4-0e9d-4ac8-b359-65111262cfad 11768 0 2022-11-26 05:05:22 +0000 UTC <nil> <nil> map[beta.kubernetes.io/arch:amd64 beta.kubernetes.io/instance-type:n1-standard-2 beta.kubernetes.io/os:linux cloud.google.com/metadata-proxy-ready:true failure-domain.beta.kubernetes.io/region:us-west1 failure-domain.beta.kubernetes.io/zone:us-west1-b kubernetes.io/arch:amd64 kubernetes.io/hostname:bootstrap-e2e-minion-group-0975 kubernetes.io/os:linux node.kubernetes.io/instance-type:n1-standard-2 topology.hostpath.csi/node:bootstrap-e2e-minion-group-0975 topology.kubernetes.io/region:us-west1 topology.kubernetes.io/zone:us-west1-b] map[csi.volume.kubernetes.io/nodeid:{"csi-hostpath-multivolume-593":"bootstrap-e2e-minion-group-0975","csi-hostpath-provisioning-7492":"bootstrap-e2e-minion-group-0975"} node.alpha.kubernetes.io/ttl:0 volumes.kubernetes.io/controller-managed-attach-detach:true] [] [] [{kubelet Update v1 2022-11-26 05:05:22 +0000 UTC FieldsV1 {"f:metadata":{"f:annotations":{".":{},"f:volumes.kubernetes.io/controller-managed-attach-detach":{}},"f:labels":{".":{},"f:beta.kubernetes.io/arch":{},"f:beta.kubernetes.io/instance-type":{},"f:beta.kubernetes.io/os":{},"f:cloud.google.com/metadata-proxy-ready":{},"f:failure-domain.beta.kubernetes.io/region":{},"f:failure-domain.beta.kubernetes.io/zone":{},"f:kubernetes.io/arch":{},"f:kubernetes.io/hostname":{},"f:kubernetes.io/os":{},"f:node.kubernetes.io/instance-type":{},"f:topology.kubernetes.io/region":{},"f:topology.kubernetes.io/zone":{}}},"f:spec":{"f:providerID":{}}} } {kube-controller-manager Update v1 2022-11-26 05:05:23 +0000 UTC FieldsV1 {"f:metadata":{"f:annotations":{"f:node.alpha.kubernetes.io/ttl":{}}},"f:spec":{"f:podCIDR":{},"f:podCIDRs":{".":{},"v:\"10.64.2.0/24\"":{}}}} } {kube-controller-manager Update v1 2022-11-26 05:13:31 +0000 UTC FieldsV1 {"f:status":{"f:conditions":{"k:{\"type\":\"NetworkUnavailable\"}":{"f:lastHeartbeatTime":{},"f:lastTransitionTime":{},"f:message":{},"f:reason":{},"f:status":{}}}}} status} {node-problem-detector Update v1 2022-11-26 05:30:31 +0000 UTC FieldsV1 {"f:status":{"f:conditions":{"k:{\"type\":\"CorruptDockerOverlay2\"}":{".":{},"f:lastHeartbeatTime":{},"f:lastTransitionTime":{},"f:message":{},"f:reason":{},"f:status":{},"f:type":{}},"k:{\"type\":\"FrequentContainerdRestart\"}":{".":{},"f:lastHeartbeatTime":{},"f:lastTransitionTime":{},"f:message":{},"f:reason":{},"f:status":{},"f:type":{}},"k:{\"type\":\"FrequentDockerRestart\"}":{".":{},"f:lastHeartbeatTime":{},"f:lastTransitionTime":{},"f:message":{},"f:reason":{},"f:status":{},"f:type":{}},"k:{\"type\":\"FrequentKubeletRestart\"}":{".":{},"f:lastHeartbeatTime":{},"f:lastTransitionTime":{},"f:message":{},"f:reason":{},"f:status":{},"f:type":{}},"k:{\"type\":\"FrequentUnregisterNetDevice\"}":{".":{},"f:lastHeartbeatTime":{},"f:lastTransitionTime":{},"f:message":{},"f:reason":{},"f:status":{},"f:type":{}},"k:{\"type\":\"KernelDeadlock\"}":{".":{},"f:lastHeartbeatTime":{},"f:lastTransitionTime":{},"f:message":{},"f:reason":{},"f:status":{},"f:type":{}},"k:{\"type\":\"ReadonlyFilesystem\"}":{".":{},"f:lastHeartbeatTime":{},"f:lastTransitionTime":{},"f:message":{},"f:reason":{},"f:status":{},"f:type":{}}}}} status} {kubelet Update v1 2022-11-26 05:30:35 +0000 UTC FieldsV1 {"f:metadata":{"f:annotations":{"f:csi.volume.kubernetes.io/nodeid":{}},"f:labels":{"f:topology.hostpath.csi/node":{}}},"f:status":{"f:conditions":{"k:{\"type\":\"DiskPressure\"}":{"f:lastHeartbeatTime":{}},"k:{\"type\":\"MemoryPressure\"}":{"f:lastHeartbeatTime":{}},"k:{\"type\":\"PIDPressure\"}":{"f:lastHeartbeatTime":{}},"k:{\"type\":\"Ready\"}":{"f:lastHeartbeatTime":{},"f:lastTransitionTime":{},"f:message":{},"f:reason":{},"f:status":{}}},"f:images":{}}} status}]},Spec:NodeSpec{PodCIDR:10.64.2.0/24,DoNotUseExternalID:,ProviderID:gce://k8s-jkns-e2e-gce-ci-reboot/us-west1-b/bootstrap-e2e-minion-group-0975,Unschedulable:false,Taints:[]Taint{},ConfigSource:nil,PodCIDRs:[10.64.2.0/24],},Status:NodeStatus{Capacity:ResourceList{attachable-volumes-gce-pd: {{127 0} {<nil>} 127 DecimalSI},cpu: {{2 0} {<nil>} 2 DecimalSI},ephemeral-storage: {{101203873792 0} {<nil>} 98831908Ki BinarySI},hugepages-1Gi: {{0 0} {<nil>} 0 DecimalSI},hugepages-2Mi: {{0 0} {<nil>} 0 DecimalSI},memory: {{7815430144 0} {<nil>} BinarySI},pods: {{110 0} {<nil>} 110 DecimalSI},},Allocatable:ResourceList{attachable-volumes-gce-pd: {{127 0} {<nil>} 127 DecimalSI},cpu: {{2 0} {<nil>} 2 DecimalSI},ephemeral-storage: {{91083486262 0} {<nil>} 91083486262 DecimalSI},hugepages-1Gi: {{0 0} {<nil>} 0 DecimalSI},hugepages-2Mi: {{0 0} {<nil>} 0 DecimalSI},memory: {{7553286144 0} {<nil>} BinarySI},pods: {{110 0} {<nil>} 110 DecimalSI},},Phase:,Conditions:[]NodeCondition{NodeCondition{Type:CorruptDockerOverlay2,Status:False,LastHeartbeatTime:2022-11-26 05:30:31 +0000 UTC,LastTransitionTime:2022-11-26 05:05:27 +0000 UTC,Reason:NoCorruptDockerOverlay2,Message:docker overlay2 is functioning properly,},NodeCondition{Type:FrequentUnregisterNetDevice,Status:False,LastHeartbeatTime:2022-11-26 05:30:31 +0000 UTC,LastTransitionTime:2022-11-26 05:05:27 +0000 UTC,Reason:NoFrequentUnregisterNetDevice,Message:node is functioning properly,},NodeCondition{Type:FrequentKubeletRestart,Status:False,LastHeartbeatTime:2022-11-26 05:30:31 +0000 UTC,LastTransitionTime:2022-11-26 05:05:27 +0000 UTC,Reason:NoFrequentKubeletRestart,Message:kubelet is functioning properly,},NodeCondition{Type:FrequentDockerRestart,Status:False,LastHeartbeatTime:2022-11-26 05:30:31 +0000 UTC,LastTransitionTime:2022-11-26 05:05:27 +0000 UTC,Reason:NoFrequentDockerRestart,Message:docker is functioning properly,},NodeCondition{Type:FrequentContainerdRestart,Status:False,LastHeartbeatTime:2022-11-26 05:30:31 +0000 UTC,LastTransitionTime:2022-11-26 05:05:27 +0000 UTC,Reason:NoFrequentContainerdRestart,Message:containerd is functioning properly,},NodeCondition{Type:KernelDeadlock,Status:False,LastHeartbeatTime:2022-11-26 05:30:31 +0000 UTC,LastTransitionTime:2022-11-26 05:05:27 +0000 UTC,Reason:KernelHasNoDeadlock,Message:kernel has no deadlock,},NodeCondition{Type:ReadonlyFilesystem,Status:False,LastHeartbeatTime:2022-11-26 05:30:31 +0000 UTC,LastTransitionTime:2022-11-26 05:05:27 +0000 UTC,Reason:FilesystemIsNotReadOnly,Message:Filesystem is not read-only,},NodeCondition{Type:NetworkUnavailable,Status:False,LastHeartbeatTime:2022-11-26 05:05:39 +0000 UTC,LastTransitionTime:2022-11-26 05:05:39 +0000 UTC,Reason:RouteCreated,Message:RouteController created a route,},NodeCondition{Type:MemoryPressure,Status:False,LastHeartbeatTime:2022-11-26 05:29:28 +0000 UTC,LastTransitionTime:2022-11-26 05:05:22 +0000 UTC,Reason:KubeletHasSufficientMemory,Message:kubelet has sufficient memory available,},NodeCondition{Type:DiskPressure,Status:False,LastHeartbeatTime:2022-11-26 05:29:28 +0000 UTC,LastTransitionTime:2022-11-26 05:05:22 +0000 UTC,Reason:KubeletHasNoDiskPressure,Message:kubelet has no disk pressure,},NodeCondition{Type:PIDPressure,Status:False,LastHeartbeatTime:2022-11-26 05:29:28 +0000 UTC,LastTransitionTime:2022-11-26 05:05:22 +0000 UTC,Reason:KubeletHasSufficientPID,Message:kubelet has sufficient PID available,},NodeCondition{Type:Ready,Status:True,LastHeartbeatTime:2022-11-26 05:29:28 +0000 UTC,LastTransitionTime:2022-11-26 05:05:23 +0000 UTC,Reason:KubeletReady,Message:kubelet is posting ready status. AppArmor enabled,},},Addresses:[]NodeAddress{NodeAddress{Type:InternalIP,Address:10.138.0.5,},NodeAddress{Type:ExternalIP,Address:35.247.104.236,},NodeAddress{Type:InternalDNS,Address:bootstrap-e2e-minion-group-0975.c.k8s-jkns-e2e-gce-ci-reboot.internal,},NodeAddress{Type:Hostname,Address:bootstrap-e2e-minion-group-0975.c.k8s-jkns-e2e-gce-ci-reboot.internal,},},DaemonEndpoints:NodeDaemonEndpoints{KubeletEndpoint:DaemonEndpoint{Port:10250,},},NodeInfo:NodeSystemInfo{MachineID:b38385f7976a474936da82f5f565760a,SystemUUID:b38385f7-976a-4749-36da-82f5f565760a,BootID:0aa74273-9b8d-48ac-bad0-0a1805f010ea,KernelVersion:5.10.123+,OSImage:Container-Optimized OS from Google,ContainerRuntimeVersion:containerd://1.7.0-beta.0-149-gd06318622,KubeletVersion:v1.27.0-alpha.0.50+70617042976dc1,KubeProxyVersion:v1.27.0-alpha.0.50+70617042976dc1,OperatingSystem:linux,Architecture:amd64,},Images:[]ContainerImage{ContainerImage{Names:[registry.k8s.io/e2e-test-images/jessie-dnsutils@sha256:24aaf2626d6b27864c29de2097e8bbb840b3a414271bf7c8995e431e47d8408e registry.k8s.io/e2e-test-images/jessie-dnsutils:1.7],SizeBytes:112030336,},ContainerImage{Names:[registry.k8s.io/e2e-test-images/volume/nfs@sha256:3bda73f2428522b0e342af80a0b9679e8594c2126f2b3cca39ed787589741b9e registry.k8s.io/e2e-test-images/volume/nfs:1.3],SizeBytes:95836203,},ContainerImage{Names:[registry.k8s.io/sig-storage/nfs-provisioner@sha256:e943bb77c7df05ebdc8c7888b2db289b13bf9f012d6a3a5a74f14d4d5743d439 registry.k8s.io/sig-storage/nfs-provisioner:v3.0.1],SizeBytes:90632047,},ContainerImage{Names:[registry.k8s.io/kube-proxy-amd64:v1.27.0-alpha.0.50_70617042976dc1],SizeBytes:67201736,},ContainerImage{Names:[registry.k8s.io/e2e-test-images/agnhost@sha256:16bbf38c463a4223d8cfe4da12bc61010b082a79b4bb003e2d3ba3ece5dd5f9e registry.k8s.io/e2e-test-images/agnhost:2.43],SizeBytes:51706353,},ContainerImage{Names:[gke.gcr.io/prometheus-to-sd@sha256:e739643c3939ba0b161425f45a1989eedfc4a3b166db9a7100863296b4c70510 gke.gcr.io/prometheus-to-sd:v0.11.1-gke.1],SizeBytes:48742566,},ContainerImage{Names:[registry.k8s.io/e2e-test-images/httpd@sha256:148b022f5c5da426fc2f3c14b5c0867e58ef05961510c84749ac1fddcb0fef22 registry.k8s.io/e2e-test-images/httpd:2.4.38-4],SizeBytes:40764257,},ContainerImage{Names:[registry.k8s.io/metrics-server/metrics-server@sha256:6385aec64bb97040a5e692947107b81e178555c7a5b71caa90d733e4130efc10 registry.k8s.io/metrics-server/metrics-server:v0.5.2],SizeBytes:26023008,},ContainerImage{Names:[registry.k8s.io/sig-storage/csi-provisioner@sha256:ee3b525d5b89db99da3b8eb521d9cd90cb6e9ef0fbb651e98bb37be78d36b5b8 registry.k8s.io/sig-storage/csi-provisioner:v3.3.0],SizeBytes:25491225,},ContainerImage{Names:[registry.k8s.io/sig-storage/csi-resizer@sha256:425d8f1b769398127767b06ed97ce62578a3179bcb99809ce93a1649e025ffe7 registry.k8s.io/sig-storage/csi-resizer:v1.6.0],SizeBytes:24148884,},ContainerImage{Names:[registry.k8s.io/sig-storage/csi-snapshotter@sha256:291334908ddf71a4661fd7f6d9d97274de8a5378a2b6fdfeb2ce73414a34f82f registry.k8s.io/sig-storage/csi-snapshotter:v6.1.0],SizeBytes:23881995,},ContainerImage{Names:[registry.k8s.io/sig-storage/csi-attacher@sha256:9a685020911e2725ad019dbce6e4a5ab93d51e3d4557f115e64343345e05781b registry.k8s.io/sig-storage/csi-attacher:v4.0.0],SizeBytes:23847201,},ContainerImage{Names:[registry.k8s.io/sig-storage/hostpathplugin@sha256:92257881c1d6493cf18299a24af42330f891166560047902b8d431fb66b01af5 registry.k8s.io/sig-storage/hostpathplugin:v1.9.0],SizeBytes:18758628,},ContainerImage{Names:[registry.k8s.io/autoscaling/addon-resizer@sha256:43f129b81d28f0fdd54de6d8e7eacd5728030782e03db16087fc241ad747d3d6 registry.k8s.io/autoscaling/addon-resizer:1.8.14],SizeBytes:10153852,},ContainerImage{Names:[registry.k8s.io/sig-storage/csi-node-driver-registrar@sha256:0103eee7c35e3e0b5cd8cdca9850dc71c793cdeb6669d8be7a89440da2d06ae4 registry.k8s.io/sig-storage/csi-node-driver-registrar:v2.5.1],SizeBytes:9133109,},ContainerImage{Names:[registry.k8s.io/sig-storage/livenessprobe@sha256:933940f13b3ea0abc62e656c1aa5c5b47c04b15d71250413a6b821bd0c58b94e registry.k8s.io/sig-storage/livenessprobe:v2.7.0],SizeBytes:8688564,},ContainerImage{Names:[registry.k8s.io/kas-network-proxy/proxy-agent@sha256:48f2a4ec3e10553a81b8dd1c6fa5fe4bcc9617f78e71c1ca89c6921335e2d7da registry.k8s.io/kas-network-proxy/proxy-agent:v0.0.33],SizeBytes:8512162,},ContainerImage{Names:[registry.k8s.io/metadata-proxy@sha256:e914645f22e946bce5165737e1b244e0a296ad1f0f81a9531adc57af2780978a registry.k8s.io/metadata-proxy:v0.1.12],SizeBytes:5301657,},ContainerImage{Names:[registry.k8s.io/e2e-test-images/busybox@sha256:c318242786b139d18676b1c09a0ad7f15fc17f8f16a5b2e625cd0dc8c9703daf registry.k8s.io/e2e-test-images/busybox:1.29-2],SizeBytes:732424,},ContainerImage{Names:[registry.k8s.io/e2e-test-images/busybox@sha256:2e0f836850e09b8b7cc937681d6194537a09fbd5f6b9e08f4d646a85128e8937 registry.k8s.io/e2e-test-images/busybox:1.29-4],SizeBytes:731990,},ContainerImage{Names:[registry.k8s.io/pause@sha256:9001185023633d17a2f98ff69b6ff2615b8ea02a825adffa40422f51dfdcde9d registry.k8s.io/pause:3.8],SizeBytes:311286,},},VolumesInUse:[],VolumesAttached:[]AttachedVolume{},Config:nil,},} Nov 26 05:32:34.270: INFO: Logging kubelet events for node bootstrap-e2e-minion-group-0975 Nov 26 05:32:34.318: INFO: Logging pods the kubelet thinks is on node bootstrap-e2e-minion-group-0975 Nov 26 05:32:34.371: INFO: Unable to retrieve kubelet pods for node bootstrap-e2e-minion-group-0975: error trying to reach service: No agent available Nov 26 05:32:34.371: INFO: Logging node info for node bootstrap-e2e-minion-group-rwsn Nov 26 05:32:34.413: INFO: Node Info: &Node{ObjectMeta:{bootstrap-e2e-minion-group-rwsn adedf225-b462-4305-a60c-9042d31ca1d2 11717 0 2022-11-26 05:05:15 +0000 UTC <nil> <nil> map[beta.kubernetes.io/arch:amd64 beta.kubernetes.io/instance-type:n1-standard-2 beta.kubernetes.io/os:linux cloud.google.com/metadata-proxy-ready:true failure-domain.beta.kubernetes.io/region:us-west1 failure-domain.beta.kubernetes.io/zone:us-west1-b kubernetes.io/arch:amd64 kubernetes.io/hostname:bootstrap-e2e-minion-group-rwsn kubernetes.io/os:linux node.kubernetes.io/instance-type:n1-standard-2 topology.hostpath.csi/node:bootstrap-e2e-minion-group-rwsn topology.kubernetes.io/region:us-west1 topology.kubernetes.io/zone:us-west1-b] map[node.alpha.kubernetes.io/ttl:0 volumes.kubernetes.io/controller-managed-attach-detach:true] [] [] [{kubelet Update v1 2022-11-26 05:05:15 +0000 UTC FieldsV1 {"f:metadata":{"f:annotations":{".":{},"f:volumes.kubernetes.io/controller-managed-attach-detach":{}},"f:labels":{".":{},"f:beta.kubernetes.io/arch":{},"f:beta.kubernetes.io/instance-type":{},"f:beta.kubernetes.io/os":{},"f:cloud.google.com/metadata-proxy-ready":{},"f:failure-domain.beta.kubernetes.io/region":{},"f:failure-domain.beta.kubernetes.io/zone":{},"f:kubernetes.io/arch":{},"f:kubernetes.io/hostname":{},"f:kubernetes.io/os":{},"f:node.kubernetes.io/instance-type":{},"f:topology.kubernetes.io/region":{},"f:topology.kubernetes.io/zone":{}}},"f:spec":{"f:providerID":{}}} } {kube-controller-manager Update v1 2022-11-26 05:05:17 +0000 UTC FieldsV1 {"f:metadata":{"f:annotations":{"f:node.alpha.kubernetes.io/ttl":{}}},"f:spec":{"f:podCIDR":{},"f:podCIDRs":{".":{},"v:\"10.64.0.0/24\"":{}}}} } {kube-controller-manager Update v1 2022-11-26 05:29:28 +0000 UTC FieldsV1 {"f:status":{"f:conditions":{"k:{\"type\":\"NetworkUnavailable\"}":{"f:lastHeartbeatTime":{},"f:lastTransitionTime":{},"f:message":{},"f:reason":{},"f:status":{}}},"f:volumesAttached":{}}} status} {kubelet Update v1 2022-11-26 05:29:47 +0000 UTC FieldsV1 {"f:metadata":{"f:labels":{"f:topology.hostpath.csi/node":{}}},"f:status":{"f:conditions":{"k:{\"type\":\"DiskPressure\"}":{"f:lastHeartbeatTime":{}},"k:{\"type\":\"MemoryPressure\"}":{"f:lastHeartbeatTime":{}},"k:{\"type\":\"PIDPressure\"}":{"f:lastHeartbeatTime":{}},"k:{\"type\":\"Ready\"}":{"f:lastHeartbeatTime":{},"f:lastTransitionTime":{},"f:message":{},"f:reason":{},"f:status":{}}},"f:images":{},"f:volumesInUse":{}}} status} {node-problem-detector Update v1 2022-11-26 05:30:22 +0000 UTC FieldsV1 {"f:status":{"f:conditions":{"k:{\"type\":\"CorruptDockerOverlay2\"}":{".":{},"f:lastHeartbeatTime":{},"f:lastTransitionTime":{},"f:message":{},"f:reason":{},"f:status":{},"f:type":{}},"k:{\"type\":\"FrequentContainerdRestart\"}":{".":{},"f:lastHeartbeatTime":{},"f:lastTransitionTime":{},"f:message":{},"f:reason":{},"f:status":{},"f:type":{}},"k:{\"type\":\"FrequentDockerRestart\"}":{".":{},"f:lastHeartbeatTime":{},"f:lastTransitionTime":{},"f:message":{},"f:reason":{},"f:status":{},"f:type":{}},"k:{\"type\":\"FrequentKubeletRestart\"}":{".":{},"f:lastHeartbeatTime":{},"f:lastTransitionTime":{},"f:message":{},"f:reason":{},"f:status":{},"f:type":{}},"k:{\"type\":\"FrequentUnregisterNetDevice\"}":{".":{},"f:lastHeartbeatTime":{},"f:lastTransitionTime":{},"f:message":{},"f:reason":{},"f:status":{},"f:type":{}},"k:{\"type\":\"KernelDeadlock\"}":{".":{},"f:lastHeartbeatTime":{},"f:lastTransitionTime":{},"f:message":{},"f:reason":{},"f:status":{},"f:type":{}},"k:{\"type\":\"ReadonlyFilesystem\"}":{".":{},"f:lastHeartbeatTime":{},"f:lastTransitionTime":{},"f:message":{},"f:reason":{},"f:status":{},"f:type":{}}}}} status}]},Spec:NodeSpec{PodCIDR:10.64.0.0/24,DoNotUseExternalID:,ProviderID:gce://k8s-jkns-e2e-gce-ci-reboot/us-west1-b/bootstrap-e2e-minion-group-rwsn,Unschedulable:false,Taints:[]Taint{},ConfigSource:nil,PodCIDRs:[10.64.0.0/24],},Status:NodeStatus{Capacity:ResourceList{attachable-volumes-gce-pd: {{127 0} {<nil>} 127 DecimalSI},cpu: {{2 0} {<nil>} 2 DecimalSI},ephemeral-storage: {{101203873792 0} {<nil>} 98831908Ki BinarySI},hugepages-1Gi: {{0 0} {<nil>} 0 DecimalSI},hugepages-2Mi: {{0 0} {<nil>} 0 DecimalSI},memory: {{7815438336 0} {<nil>} BinarySI},pods: {{110 0} {<nil>} 110 DecimalSI},},Allocatable:ResourceList{attachable-volumes-gce-pd: {{127 0} {<nil>} 127 DecimalSI},cpu: {{2 0} {<nil>} 2 DecimalSI},ephemeral-storage: {{91083486262 0} {<nil>} 91083486262 DecimalSI},hugepages-1Gi: {{0 0} {<nil>} 0 DecimalSI},hugepages-2Mi: {{0 0} {<nil>} 0 DecimalSI},memory: {{7553294336 0} {<nil>} BinarySI},pods: {{110 0} {<nil>} 110 DecimalSI},},Phase:,Conditions:[]NodeCondition{NodeCondition{Type:FrequentContainerdRestart,Status:False,LastHeartbeatTime:2022-11-26 05:30:22 +0000 UTC,LastTransitionTime:2022-11-26 05:05:18 +0000 UTC,Reason:NoFrequentContainerdRestart,Message:containerd is functioning properly,},NodeCondition{Type:KernelDeadlock,Status:False,LastHeartbeatTime:2022-11-26 05:30:22 +0000 UTC,LastTransitionTime:2022-11-26 05:05:18 +0000 UTC,Reason:KernelHasNoDeadlock,Message:kernel has no deadlock,},NodeCondition{Type:ReadonlyFilesystem,Status:False,LastHeartbeatTime:2022-11-26 05:30:22 +0000 UTC,LastTransitionTime:2022-11-26 05:05:18 +0000 UTC,Reason:FilesystemIsNotReadOnly,Message:Filesystem is not read-only,},NodeCondition{Type:CorruptDockerOverlay2,Status:False,LastHeartbeatTime:2022-11-26 05:30:22 +0000 UTC,LastTransitionTime:2022-11-26 05:05:18 +0000 UTC,Reason:NoCorruptDockerOverlay2,Message:docker overlay2 is functioning properly,},NodeCondition{Type:FrequentUnregisterNetDevice,Status:False,LastHeartbeatTime:2022-11-26 05:30:22 +0000 UTC,LastTransitionTime:2022-11-26 05:05:18 +0000 UTC,Reason:NoFrequentUnregisterNetDevice,Message:node is functioning properly,},NodeCondition{Type:FrequentKubeletRestart,Status:False,LastHeartbeatTime:2022-11-26 05:30:22 +0000 UTC,LastTransitionTime:2022-11-26 05:05:18 +0000 UTC,Reason:NoFrequentKubeletRestart,Message:kubelet is functioning properly,},NodeCondition{Type:FrequentDockerRestart,Status:False,LastHeartbeatTime:2022-11-26 05:30:22 +0000 UTC,LastTransitionTime:2022-11-26 05:05:18 +0000 UTC,Reason:NoFrequentDockerRestart,Message:docker is functioning properly,},NodeCondition{Type:NetworkUnavailable,Status:False,LastHeartbeatTime:2022-11-26 05:05:25 +0000 UTC,LastTransitionTime:2022-11-26 05:05:25 +0000 UTC,Reason:RouteCreated,Message:RouteController created a route,},NodeCondition{Type:MemoryPressure,Status:False,LastHeartbeatTime:2022-11-26 05:27:50 +0000 UTC,LastTransitionTime:2022-11-26 05:05:15 +0000 UTC,Reason:KubeletHasSufficientMemory,Message:kubelet has sufficient memory available,},NodeCondition{Type:DiskPressure,Status:False,LastHeartbeatTime:2022-11-26 05:27:50 +0000 UTC,LastTransitionTime:2022-11-26 05:05:15 +0000 UTC,Reason:KubeletHasNoDiskPressure,Message:kubelet has no disk pressure,},NodeCondition{Type:PIDPressure,Status:False,LastHeartbeatTime:2022-11-26 05:27:50 +0000 UTC,LastTransitionTime:2022-11-26 05:05:15 +0000 UTC,Reason:KubeletHasSufficientPID,Message:kubelet has sufficient PID available,},NodeCondition{Type:Ready,Status:True,LastHeartbeatTime:2022-11-26 05:27:50 +0000 UTC,LastTransitionTime:2022-11-26 05:05:17 +0000 UTC,Reason:KubeletReady,Message:kubelet is posting ready status. AppArmor enabled,},},Addresses:[]NodeAddress{NodeAddress{Type:InternalIP,Address:10.138.0.4,},NodeAddress{Type:ExternalIP,Address:35.230.27.45,},NodeAddress{Type:InternalDNS,Address:bootstrap-e2e-minion-group-rwsn.c.k8s-jkns-e2e-gce-ci-reboot.internal,},NodeAddress{Type:Hostname,Address:bootstrap-e2e-minion-group-rwsn.c.k8s-jkns-e2e-gce-ci-reboot.internal,},},DaemonEndpoints:NodeDaemonEndpoints{KubeletEndpoint:DaemonEndpoint{Port:10250,},},NodeInfo:NodeSystemInfo{MachineID:69e22ce1281686d6cde1b3211b34b8af,SystemUUID:69e22ce1-2816-86d6-cde1-b3211b34b8af,BootID:79e9e1fd-2a39-42ec-b72b-72836d213862,KernelVersion:5.10.123+,OSImage:Container-Optimized OS from Google,ContainerRuntimeVersion:containerd://1.7.0-beta.0-149-gd06318622,KubeletVersion:v1.27.0-alpha.0.50+70617042976dc1,KubeProxyVersion:v1.27.0-alpha.0.50+70617042976dc1,OperatingSystem:linux,Architecture:amd64,},Images:[]ContainerImage{ContainerImage{Names:[registry.k8s.io/e2e-test-images/jessie-dnsutils@sha256:24aaf2626d6b27864c29de2097e8bbb840b3a414271bf7c8995e431e47d8408e registry.k8s.io/e2e-test-images/jessie-dnsutils:1.7],SizeBytes:112030336,},ContainerImage{Names:[registry.k8s.io/e2e-test-images/volume/nfs@sha256:3bda73f2428522b0e342af80a0b9679e8594c2126f2b3cca39ed787589741b9e registry.k8s.io/e2e-test-images/volume/nfs:1.3],SizeBytes:95836203,},ContainerImage{Names:[registry.k8s.io/sig-storage/nfs-provisioner@sha256:e943bb77c7df05ebdc8c7888b2db289b13bf9f012d6a3a5a74f14d4d5743d439 registry.k8s.io/sig-storage/nfs-provisioner:v3.0.1],SizeBytes:90632047,},ContainerImage{Names:[registry.k8s.io/kube-proxy-amd64:v1.27.0-alpha.0.50_70617042976dc1],SizeBytes:67201736,},ContainerImage{Names:[registry.k8s.io/e2e-test-images/agnhost@sha256:16bbf38c463a4223d8cfe4da12bc61010b082a79b4bb003e2d3ba3ece5dd5f9e registry.k8s.io/e2e-test-images/agnhost:2.43],SizeBytes:51706353,},ContainerImage{Names:[gke.gcr.io/prometheus-to-sd@sha256:e739643c3939ba0b161425f45a1989eedfc4a3b166db9a7100863296b4c70510 gke.gcr.io/prometheus-to-sd:v0.11.1-gke.1],SizeBytes:48742566,},ContainerImage{Names:[registry.k8s.io/metrics-server/metrics-server@sha256:6385aec64bb97040a5e692947107b81e178555c7a5b71caa90d733e4130efc10 registry.k8s.io/metrics-server/metrics-server:v0.5.2],SizeBytes:26023008,},ContainerImage{Names:[registry.k8s.io/sig-storage/csi-provisioner@sha256:ee3b525d5b89db99da3b8eb521d9cd90cb6e9ef0fbb651e98bb37be78d36b5b8 registry.k8s.io/sig-storage/csi-provisioner:v3.3.0],SizeBytes:25491225,},ContainerImage{Names:[registry.k8s.io/sig-storage/csi-resizer@sha256:425d8f1b769398127767b06ed97ce62578a3179bcb99809ce93a1649e025ffe7 registry.k8s.io/sig-storage/csi-resizer:v1.6.0],SizeBytes:24148884,},ContainerImage{Names:[registry.k8s.io/sig-storage/csi-snapshotter@sha256:291334908ddf71a4661fd7f6d9d97274de8a5378a2b6fdfeb2ce73414a34f82f registry.k8s.io/sig-storage/csi-snapshotter:v6.1.0],SizeBytes:23881995,},ContainerImage{Names:[registry.k8s.io/sig-storage/csi-attacher@sha256:9a685020911e2725ad019dbce6e4a5ab93d51e3d4557f115e64343345e05781b registry.k8s.io/sig-storage/csi-attacher:v4.0.0],SizeBytes:23847201,},ContainerImage{Names:[registry.k8s.io/sig-storage/snapshot-controller@sha256:823c75d0c45d1427f6d850070956d9ca657140a7bbf828381541d1d808475280 registry.k8s.io/sig-storage/snapshot-controller:v6.1.0],SizeBytes:22620891,},ContainerImage{Names:[registry.k8s.io/sig-storage/hostpathplugin@sha256:92257881c1d6493cf18299a24af42330f891166560047902b8d431fb66b01af5 registry.k8s.io/sig-storage/hostpathplugin:v1.9.0],SizeBytes:18758628,},ContainerImage{Names:[registry.k8s.io/cpa/cluster-proportional-autoscaler@sha256:fd636b33485c7826fb20ef0688a83ee0910317dbb6c0c6f3ad14661c1db25def registry.k8s.io/cpa/cluster-proportional-autoscaler:1.8.4],SizeBytes:15209393,},ContainerImage{Names:[registry.k8s.io/coredns/coredns@sha256:8e352a029d304ca7431c6507b56800636c321cb52289686a581ab70aaa8a2e2a registry.k8s.io/coredns/coredns:v1.9.3],SizeBytes:14837849,},ContainerImage{Names:[registry.k8s.io/autoscaling/addon-resizer@sha256:43f129b81d28f0fdd54de6d8e7eacd5728030782e03db16087fc241ad747d3d6 registry.k8s.io/autoscaling/addon-resizer:1.8.14],SizeBytes:10153852,},ContainerImage{Names:[registry.k8s.io/sig-storage/csi-node-driver-registrar@sha256:0103eee7c35e3e0b5cd8cdca9850dc71c793cdeb6669d8be7a89440da2d06ae4 registry.k8s.io/sig-storage/csi-node-driver-registrar:v2.5.1],SizeBytes:9133109,},ContainerImage{Names:[registry.k8s.io/sig-storage/livenessprobe@sha256:933940f13b3ea0abc62e656c1aa5c5b47c04b15d71250413a6b821bd0c58b94e registry.k8s.io/sig-storage/livenessprobe:v2.7.0],SizeBytes:8688564,},ContainerImage{Names:[registry.k8s.io/kas-network-proxy/proxy-agent@sha256:48f2a4ec3e10553a81b8dd1c6fa5fe4bcc9617f78e71c1ca89c6921335e2d7da registry.k8s.io/kas-network-proxy/proxy-agent:v0.0.33],SizeBytes:8512162,},ContainerImage{Names:[registry.k8s.io/networking/ingress-gce-404-server-with-metrics-amd64@sha256:7eb7b3cee4d33c10c49893ad3c386232b86d4067de5251294d4c620d6e072b93 registry.k8s.io/networking/ingress-gce-404-server-with-metrics-amd64:v1.10.11],SizeBytes:6463068,},ContainerImage{Names:[registry.k8s.io/metadata-proxy@sha256:e914645f22e946bce5165737e1b244e0a296ad1f0f81a9531adc57af2780978a registry.k8s.io/metadata-proxy:v0.1.12],SizeBytes:5301657,},ContainerImage{Names:[registry.k8s.io/e2e-test-images/busybox@sha256:c318242786b139d18676b1c09a0ad7f15fc17f8f16a5b2e625cd0dc8c9703daf registry.k8s.io/e2e-test-images/busybox:1.29-2],SizeBytes:732424,},ContainerImage{Names:[registry.k8s.io/e2e-test-images/busybox@sha256:2e0f836850e09b8b7cc937681d6194537a09fbd5f6b9e08f4d646a85128e8937 registry.k8s.io/e2e-test-images/busybox:1.29-4],SizeBytes:731990,},ContainerImage{Names:[registry.k8s.io/pause@sha256:7031c1b283388d2c2e09b57badb803c05ebed362dc88d84b480cc47f72a21097 registry.k8s.io/pause:3.9],SizeBytes:321520,},ContainerImage{Names:[registry.k8s.io/pause@sha256:9001185023633d17a2f98ff69b6ff2615b8ea02a825adffa40422f51dfdcde9d registry.k8s.io/pause:3.8],SizeBytes:311286,},},VolumesInUse:[kubernetes.io/csi/csi-hostpath-multivolume-4683^6fe7a3e3-6d49-11ed-ac34-8a6e303c97ed kubernetes.io/csi/csi-hostpath-multivolume-4683^7ad9a9cd-6d49-11ed-b2a0-f2f9c395d3b2 kubernetes.io/csi/csi-hostpath-multivolume-6872^8dfb6ea6-6d49-11ed-bbbd-7a3847645704 kubernetes.io/csi/csi-hostpath-provisioning-6984^3b0457df-6d48-11ed-9e4d-d626cdd9a50a],VolumesAttached:[]AttachedVolume{AttachedVolume{Name:kubernetes.io/csi/csi-hostpath-multivolume-6872^8dfb6ea6-6d49-11ed-bbbd-7a3847645704,DevicePath:,},AttachedVolume{Name:kubernetes.io/csi/csi-hostpath-provisioning-6984^3b0457df-6d48-11ed-9e4d-d626cdd9a50a,DevicePath:,},AttachedVolume{Name:kubernetes.io/csi/csi-hostpath-multivolume-4683^7ad9a9cd-6d49-11ed-b2a0-f2f9c395d3b2,DevicePath:,},AttachedVolume{Name:kubernetes.io/csi/csi-hostpath-multivolume-4683^6fe7a3e3-6d49-11ed-ac34-8a6e303c97ed,DevicePath:,},},Config:nil,},} Nov 26 05:32:34.414: INFO: Logging kubelet events for node bootstrap-e2e-minion-group-rwsn Nov 26 05:32:34.466: INFO: Logging pods the kubelet thinks is on node bootstrap-e2e-minion-group-rwsn Nov 26 05:32:34.588: INFO: Unable to retrieve kubelet pods for node bootstrap-e2e-minion-group-rwsn: error trying to reach service: No agent available Nov 26 05:32:34.588: INFO: Logging node info for node bootstrap-e2e-minion-group-vhdj Nov 26 05:32:34.632: INFO: Node Info: &Node{ObjectMeta:{bootstrap-e2e-minion-group-vhdj 0d27fd16-9ec8-4428-bb1e-97fead6c7a66 11827 0 2022-11-26 05:05:23 +0000 UTC <nil> <nil> map[beta.kubernetes.io/arch:amd64 beta.kubernetes.io/instance-type:n1-standard-2 beta.kubernetes.io/os:linux cloud.google.com/metadata-proxy-ready:true failure-domain.beta.kubernetes.io/region:us-west1 failure-domain.beta.kubernetes.io/zone:us-west1-b kubernetes.io/arch:amd64 kubernetes.io/hostname:bootstrap-e2e-minion-group-vhdj kubernetes.io/os:linux node.kubernetes.io/instance-type:n1-standard-2 topology.hostpath.csi/node:bootstrap-e2e-minion-group-vhdj topology.kubernetes.io/region:us-west1 topology.kubernetes.io/zone:us-west1-b] map[csi.volume.kubernetes.io/nodeid:{"csi-hostpath-provisioning-7617":"bootstrap-e2e-minion-group-vhdj","csi-hostpath-volumeio-4786":"bootstrap-e2e-minion-group-vhdj"} node.alpha.kubernetes.io/ttl:0 volumes.kubernetes.io/controller-managed-attach-detach:true] [] [] [{kubelet Update v1 2022-11-26 05:05:23 +0000 UTC FieldsV1 {"f:metadata":{"f:annotations":{".":{},"f:volumes.kubernetes.io/controller-managed-attach-detach":{}},"f:labels":{".":{},"f:beta.kubernetes.io/arch":{},"f:beta.kubernetes.io/instance-type":{},"f:beta.kubernetes.io/os":{},"f:cloud.google.com/metadata-proxy-ready":{},"f:failure-domain.beta.kubernetes.io/region":{},"f:failure-domain.beta.kubernetes.io/zone":{},"f:kubernetes.io/arch":{},"f:kubernetes.io/hostname":{},"f:kubernetes.io/os":{},"f:node.kubernetes.io/instance-type":{},"f:topology.kubernetes.io/region":{},"f:topology.kubernetes.io/zone":{}}},"f:spec":{"f:providerID":{}}} } {kube-controller-manager Update v1 2022-11-26 05:05:24 +0000 UTC FieldsV1 {"f:metadata":{"f:annotations":{"f:node.alpha.kubernetes.io/ttl":{}}},"f:spec":{"f:podCIDR":{},"f:podCIDRs":{".":{},"v:\"10.64.3.0/24\"":{}}}} } {kube-controller-manager Update v1 2022-11-26 05:17:38 +0000 UTC FieldsV1 {"f:status":{"f:conditions":{"k:{\"type\":\"NetworkUnavailable\"}":{"f:lastHeartbeatTime":{},"f:lastTransitionTime":{},"f:message":{},"f:reason":{},"f:status":{}}}}} status} {node-problem-detector Update v1 2022-11-26 05:30:29 +0000 UTC FieldsV1 {"f:status":{"f:conditions":{"k:{\"type\":\"CorruptDockerOverlay2\"}":{".":{},"f:lastHeartbeatTime":{},"f:lastTransitionTime":{},"f:message":{},"f:reason":{},"f:status":{},"f:type":{}},"k:{\"type\":\"FrequentContainerdRestart\"}":{".":{},"f:lastHeartbeatTime":{},"f:lastTransitionTime":{},"f:message":{},"f:reason":{},"f:status":{},"f:type":{}},"k:{\"type\":\"FrequentDockerRestart\"}":{".":{},"f:lastHeartbeatTime":{},"f:lastTransitionTime":{},"f:message":{},"f:reason":{},"f:status":{},"f:type":{}},"k:{\"type\":\"FrequentKubeletRestart\"}":{".":{},"f:lastHeartbeatTime":{},"f:lastTransitionTime":{},"f:message":{},"f:reason":{},"f:status":{},"f:type":{}},"k:{\"type\":\"FrequentUnregisterNetDevice\"}":{".":{},"f:lastHeartbeatTime":{},"f:lastTransitionTime":{},"f:message":{},"f:reason":{},"f:status":{},"f:type":{}},"k:{\"type\":\"KernelDeadlock\"}":{".":{},"f:lastHeartbeatTime":{},"f:lastTransitionTime":{},"f:message":{},"f:reason":{},"f:status":{},"f:type":{}},"k:{\"type\":\"ReadonlyFilesystem\"}":{".":{},"f:lastHeartbeatTime":{},"f:lastTransitionTime":{},"f:message":{},"f:reason":{},"f:status":{},"f:type":{}}}}} status} {kubelet Update v1 2022-11-26 05:31:02 +0000 UTC FieldsV1 {"f:metadata":{"f:annotations":{"f:csi.volume.kubernetes.io/nodeid":{}},"f:labels":{"f:topology.hostpath.csi/node":{}}},"f:status":{"f:conditions":{"k:{\"type\":\"DiskPressure\"}":{"f:lastHeartbeatTime":{}},"k:{\"type\":\"MemoryPressure\"}":{"f:lastHeartbeatTime":{}},"k:{\"type\":\"PIDPressure\"}":{"f:lastHeartbeatTime":{}},"k:{\"type\":\"Ready\"}":{"f:lastHeartbeatTime":{},"f:lastTransitionTime":{},"f:message":{},"f:reason":{},"f:status":{}}},"f:images":{}}} status}]},Spec:NodeSpec{PodCIDR:10.64.3.0/24,DoNotUseExternalID:,ProviderID:gce://k8s-jkns-e2e-gce-ci-reboot/us-west1-b/bootstrap-e2e-minion-group-vhdj,Unschedulable:false,Taints:[]Taint{},ConfigSource:nil,PodCIDRs:[10.64.3.0/24],},Status:NodeStatus{Capacity:ResourceList{attachable-volumes-gce-pd: {{127 0} {<nil>} 127 DecimalSI},cpu: {{2 0} {<nil>} 2 DecimalSI},ephemeral-storage: {{101203873792 0} {<nil>} 98831908Ki BinarySI},hugepages-1Gi: {{0 0} {<nil>} 0 DecimalSI},hugepages-2Mi: {{0 0} {<nil>} 0 DecimalSI},memory: {{7815438336 0} {<nil>} BinarySI},pods: {{110 0} {<nil>} 110 DecimalSI},},Allocatable:ResourceList{attachable-volumes-gce-pd: {{127 0} {<nil>} 127 DecimalSI},cpu: {{2 0} {<nil>} 2 DecimalSI},ephemeral-storage: {{91083486262 0} {<nil>} 91083486262 DecimalSI},hugepages-1Gi: {{0 0} {<nil>} 0 DecimalSI},hugepages-2Mi: {{0 0} {<nil>} 0 DecimalSI},memory: {{7553294336 0} {<nil>} BinarySI},pods: {{110 0} {<nil>} 110 DecimalSI},},Phase:,Conditions:[]NodeCondition{NodeCondition{Type:KernelDeadlock,Status:False,LastHeartbeatTime:2022-11-26 05:30:29 +0000 UTC,LastTransitionTime:2022-11-26 05:05:26 +0000 UTC,Reason:KernelHasNoDeadlock,Message:kernel has no deadlock,},NodeCondition{Type:ReadonlyFilesystem,Status:False,LastHeartbeatTime:2022-11-26 05:30:29 +0000 UTC,LastTransitionTime:2022-11-26 05:05:26 +0000 UTC,Reason:FilesystemIsNotReadOnly,Message:Filesystem is not read-only,},NodeCondition{Type:CorruptDockerOverlay2,Status:False,LastHeartbeatTime:2022-11-26 05:30:29 +0000 UTC,LastTransitionTime:2022-11-26 05:05:26 +0000 UTC,Reason:NoCorruptDockerOverlay2,Message:docker overlay2 is functioning properly,},NodeCondition{Type:FrequentUnregisterNetDevice,Status:False,LastHeartbeatTime:2022-11-26 05:30:29 +0000 UTC,LastTransitionTime:2022-11-26 05:05:26 +0000 UTC,Reason:NoFrequentUnregisterNetDevice,Message:node is functioning properly,},NodeCondition{Type:FrequentKubeletRestart,Status:False,LastHeartbeatTime:2022-11-26 05:30:29 +0000 UTC,LastTransitionTime:2022-11-26 05:05:26 +0000 UTC,Reason:NoFrequentKubeletRestart,Message:kubelet is functioning properly,},NodeCondition{Type:FrequentDockerRestart,Status:False,LastHeartbeatTime:2022-11-26 05:30:29 +0000 UTC,LastTransitionTime:2022-11-26 05:05:26 +0000 UTC,Reason:NoFrequentDockerRestart,Message:docker is functioning properly,},NodeCondition{Type:FrequentContainerdRestart,Status:False,LastHeartbeatTime:2022-11-26 05:30:29 +0000 UTC,LastTransitionTime:2022-11-26 05:05:26 +0000 UTC,Reason:NoFrequentContainerdRestart,Message:containerd is functioning properly,},NodeCondition{Type:NetworkUnavailable,Status:False,LastHeartbeatTime:2022-11-26 05:05:39 +0000 UTC,LastTransitionTime:2022-11-26 05:05:39 +0000 UTC,Reason:RouteCreated,Message:RouteController created a route,},NodeCondition{Type:MemoryPressure,Status:False,LastHeartbeatTime:2022-11-26 05:27:51 +0000 UTC,LastTransitionTime:2022-11-26 05:05:23 +0000 UTC,Reason:KubeletHasSufficientMemory,Message:kubelet has sufficient memory available,},NodeCondition{Type:DiskPressure,Status:False,LastHeartbeatTime:2022-11-26 05:27:51 +0000 UTC,LastTransitionTime:2022-11-26 05:05:23 +0000 UTC,Reason:KubeletHasNoDiskPressure,Message:kubelet has no disk pressure,},NodeCondition{Type:PIDPressure,Status:False,LastHeartbeatTime:2022-11-26 05:27:51 +0000 UTC,LastTransitionTime:2022-11-26 05:05:23 +0000 UTC,Reason:KubeletHasSufficientPID,Message:kubelet has sufficient PID available,},NodeCondition{Type:Ready,Status:True,LastHeartbeatTime:2022-11-26 05:27:51 +0000 UTC,LastTransitionTime:2022-11-26 05:05:24 +0000 UTC,Reason:KubeletReady,Message:kubelet is posting ready status. AppArmor enabled,},},Addresses:[]NodeAddress{NodeAddress{Type:InternalIP,Address:10.138.0.3,},NodeAddress{Type:ExternalIP,Address:34.83.94.215,},NodeAddress{Type:InternalDNS,Address:bootstrap-e2e-minion-group-vhdj.c.k8s-jkns-e2e-gce-ci-reboot.internal,},NodeAddress{Type:Hostname,Address:bootstrap-e2e-minion-group-vhdj.c.k8s-jkns-e2e-gce-ci-reboot.internal,},},DaemonEndpoints:NodeDaemonEndpoints{KubeletEndpoint:DaemonEndpoint{Port:10250,},},NodeInfo:NodeSystemInfo{MachineID:31b0953ad3e740aeb7fc1a89edb5435a,SystemUUID:31b0953a-d3e7-40ae-b7fc-1a89edb5435a,BootID:14c054d0-bda9-4ddf-b9f6-e1742be33d25,KernelVersion:5.10.123+,OSImage:Container-Optimized OS from Google,ContainerRuntimeVersion:containerd://1.7.0-beta.0-149-gd06318622,KubeletVersion:v1.27.0-alpha.0.50+70617042976dc1,KubeProxyVersion:v1.27.0-alpha.0.50+70617042976dc1,OperatingSystem:linux,Architecture:amd64,},Images:[]ContainerImage{ContainerImage{Names:[registry.k8s.io/e2e-test-images/jessie-dnsutils@sha256:24aaf2626d6b27864c29de2097e8bbb840b3a414271bf7c8995e431e47d8408e registry.k8s.io/e2e-test-images/jessie-dnsutils:1.7],SizeBytes:112030336,},ContainerImage{Names:[registry.k8s.io/sig-storage/nfs-provisioner@sha256:e943bb77c7df05ebdc8c7888b2db289b13bf9f012d6a3a5a74f14d4d5743d439 registry.k8s.io/sig-storage/nfs-provisioner:v3.0.1],SizeBytes:90632047,},ContainerImage{Names:[registry.k8s.io/kube-proxy-amd64:v1.27.0-alpha.0.50_70617042976dc1],SizeBytes:67201736,},ContainerImage{Names:[registry.k8s.io/e2e-test-images/agnhost@sha256:16bbf38c463a4223d8cfe4da12bc61010b082a79b4bb003e2d3ba3ece5dd5f9e registry.k8s.io/e2e-test-images/agnhost:2.43],SizeBytes:51706353,},ContainerImage{Names:[gke.gcr.io/prometheus-to-sd@sha256:e739643c3939ba0b161425f45a1989eedfc4a3b166db9a7100863296b4c70510 gke.gcr.io/prometheus-to-sd:v0.11.1-gke.1],SizeBytes:48742566,},ContainerImage{Names:[registry.k8s.io/e2e-test-images/httpd@sha256:148b022f5c5da426fc2f3c14b5c0867e58ef05961510c84749ac1fddcb0fef22 registry.k8s.io/e2e-test-images/httpd:2.4.38-4],SizeBytes:40764257,},ContainerImage{Names:[registry.k8s.io/sig-storage/csi-provisioner@sha256:ee3b525d5b89db99da3b8eb521d9cd90cb6e9ef0fbb651e98bb37be78d36b5b8 registry.k8s.io/sig-storage/csi-provisioner:v3.3.0],SizeBytes:25491225,},ContainerImage{Names:[registry.k8s.io/sig-storage/csi-resizer@sha256:425d8f1b769398127767b06ed97ce62578a3179bcb99809ce93a1649e025ffe7 registry.k8s.io/sig-storage/csi-resizer:v1.6.0],SizeBytes:24148884,},ContainerImage{Names:[registry.k8s.io/sig-storage/csi-snapshotter@sha256:291334908ddf71a4661fd7f6d9d97274de8a5378a2b6fdfeb2ce73414a34f82f registry.k8s.io/sig-storage/csi-snapshotter:v6.1.0],SizeBytes:23881995,},ContainerImage{Names:[registry.k8s.io/sig-storage/csi-attacher@sha256:9a685020911e2725ad019dbce6e4a5ab93d51e3d4557f115e64343345e05781b registry.k8s.io/sig-storage/csi-attacher:v4.0.0],SizeBytes:23847201,},ContainerImage{Names:[registry.k8s.io/sig-storage/hostpathplugin@sha256:92257881c1d6493cf18299a24af42330f891166560047902b8d431fb66b01af5 registry.k8s.io/sig-storage/hostpathplugin:v1.9.0],SizeBytes:18758628,},ContainerImage{Names:[registry.k8s.io/sig-storage/csi-node-driver-registrar@sha256:0103eee7c35e3e0b5cd8cdca9850dc71c793cdeb6669d8be7a89440da2d06ae4 registry.k8s.io/sig-storage/csi-node-driver-registrar:v2.5.1],SizeBytes:9133109,},ContainerImage{Names:[registry.k8s.io/sig-storage/livenessprobe@sha256:933940f13b3ea0abc62e656c1aa5c5b47c04b15d71250413a6b821bd0c58b94e registry.k8s.io/sig-storage/livenessprobe:v2.7.0],SizeBytes:8688564,},ContainerImage{Names:[registry.k8s.io/kas-network-proxy/proxy-agent@sha256:48f2a4ec3e10553a81b8dd1c6fa5fe4bcc9617f78e71c1ca89c6921335e2d7da registry.k8s.io/kas-network-proxy/proxy-agent:v0.0.33],SizeBytes:8512162,},ContainerImage{Names:[registry.k8s.io/metadata-proxy@sha256:e914645f22e946bce5165737e1b244e0a296ad1f0f81a9531adc57af2780978a registry.k8s.io/metadata-proxy:v0.1.12],SizeBytes:5301657,},ContainerImage{Names:[registry.k8s.io/e2e-test-images/busybox@sha256:2e0f836850e09b8b7cc937681d6194537a09fbd5f6b9e08f4d646a85128e8937 registry.k8s.io/e2e-test-images/busybox:1.29-4],SizeBytes:731990,},ContainerImage{Names:[registry.k8s.io/pause@sha256:9001185023633d17a2f98ff69b6ff2615b8ea02a825adffa40422f51dfdcde9d registry.k8s.io/pause:3.8],SizeBytes:311286,},},VolumesInUse:[],VolumesAttached:[]AttachedVolume{},Config:nil,},} Nov 26 05:32:34.633: INFO: Logging kubelet events for node bootstrap-e2e-minion-group-vhdj Nov 26 05:32:34.694: INFO: Logging pods the kubelet thinks is on node bootstrap-e2e-minion-group-vhdj Nov 26 05:32:34.739: INFO: Unable to retrieve kubelet pods for node bootstrap-e2e-minion-group-vhdj: error trying to reach service: No agent available [DeferCleanup (Each)] [sig-network] LoadBalancers ESIPP [Slow] tear down framework | framework.go:193 STEP: Destroying namespace "esipp-2671" for this suite. 11/26/22 05:32:34.739
Filter through log files | View test history on testgrid
go run hack/e2e.go -v --test --test_args='--ginkgo.focus=Kubernetes\se2e\ssuite\s\[It\]\s\[sig\-network\]\sLoadBalancers\sESIPP\s\[Slow\]\sshould\swork\sfor\stype\=NodePort$'
test/e2e/framework/network/utils.go:866 k8s.io/kubernetes/test/e2e/framework/network.(*NetworkingTestConfig).createNetProxyPods(0xc0010147e0, {0x75c6f7c, 0x9}, 0xc0022e0d80) test/e2e/framework/network/utils.go:866 +0x1d0 k8s.io/kubernetes/test/e2e/framework/network.(*NetworkingTestConfig).setupCore(0xc0010147e0, 0x7f18e001c5e0?) test/e2e/framework/network/utils.go:763 +0x55 k8s.io/kubernetes/test/e2e/framework/network.(*NetworkingTestConfig).setup(0xc0010147e0, 0x3c?) test/e2e/framework/network/utils.go:778 +0x3e k8s.io/kubernetes/test/e2e/framework/network.NewNetworkingTestConfig(0xc000cc0000, {0x0, 0x0, 0x0?}) test/e2e/framework/network/utils.go:131 +0x125 k8s.io/kubernetes/test/e2e/network.glob..func20.4() test/e2e/network/loadbalancer.go:1332 +0x145from junit_01.xml
[BeforeEach] [sig-network] LoadBalancers ESIPP [Slow] set up framework | framework.go:178 STEP: Creating a kubernetes client 11/26/22 05:16:00.007 Nov 26 05:16:00.007: INFO: >>> kubeConfig: /workspace/.kube/config STEP: Building a namespace api object, basename esipp 11/26/22 05:16:00.009 STEP: Waiting for a default service account to be provisioned in namespace 11/26/22 05:16:00.133 STEP: Waiting for kube-root-ca.crt to be provisioned in namespace 11/26/22 05:16:00.215 [BeforeEach] [sig-network] LoadBalancers ESIPP [Slow] test/e2e/framework/metrics/init/init.go:31 [BeforeEach] [sig-network] LoadBalancers ESIPP [Slow] test/e2e/network/loadbalancer.go:1250 [It] should work for type=NodePort test/e2e/network/loadbalancer.go:1314 STEP: creating a service esipp-5928/external-local-nodeport with type=NodePort and ExternalTrafficPolicy=Local 11/26/22 05:16:02.976 STEP: creating a pod to be part of the service external-local-nodeport 11/26/22 05:16:03.074 Nov 26 05:16:03.237: INFO: Waiting up to 2m0s for 1 pods to be created Nov 26 05:16:03.425: INFO: Found 0/1 pods - will retry Nov 26 05:16:05.481: INFO: Found all 1 pods Nov 26 05:16:05.481: INFO: Waiting up to 2m0s for 1 pods to be running and ready: [external-local-nodeport-k2k2v] Nov 26 05:16:05.481: INFO: Waiting up to 2m0s for pod "external-local-nodeport-k2k2v" in namespace "esipp-5928" to be "running and ready" Nov 26 05:16:05.549: INFO: Pod "external-local-nodeport-k2k2v": Phase="Pending", Reason="", readiness=false. Elapsed: 67.439219ms Nov 26 05:16:05.549: INFO: Error evaluating pod condition running and ready: want pod 'external-local-nodeport-k2k2v' on 'bootstrap-e2e-minion-group-vhdj' to be 'Running' but was 'Pending' Nov 26 05:16:07.701: INFO: Pod "external-local-nodeport-k2k2v": Phase="Pending", Reason="", readiness=false. Elapsed: 2.219305007s Nov 26 05:16:07.701: INFO: Error evaluating pod condition running and ready: want pod 'external-local-nodeport-k2k2v' on 'bootstrap-e2e-minion-group-vhdj' to be 'Running' but was 'Pending' Nov 26 05:16:09.606: INFO: Pod "external-local-nodeport-k2k2v": Phase="Pending", Reason="", readiness=false. Elapsed: 4.124716903s Nov 26 05:16:09.606: INFO: Error evaluating pod condition running and ready: want pod 'external-local-nodeport-k2k2v' on 'bootstrap-e2e-minion-group-vhdj' to be 'Running' but was 'Pending' Nov 26 05:16:11.601: INFO: Pod "external-local-nodeport-k2k2v": Phase="Pending", Reason="", readiness=false. Elapsed: 6.119217897s Nov 26 05:16:11.601: INFO: Error evaluating pod condition running and ready: want pod 'external-local-nodeport-k2k2v' on 'bootstrap-e2e-minion-group-vhdj' to be 'Running' but was 'Pending' Nov 26 05:16:13.802: INFO: Pod "external-local-nodeport-k2k2v": Phase="Pending", Reason="", readiness=false. Elapsed: 8.320104297s Nov 26 05:16:13.802: INFO: Error evaluating pod condition running and ready: want pod 'external-local-nodeport-k2k2v' on 'bootstrap-e2e-minion-group-vhdj' to be 'Running' but was 'Pending' Nov 26 05:16:15.599: INFO: Pod "external-local-nodeport-k2k2v": Phase="Pending", Reason="", readiness=false. Elapsed: 10.117305502s Nov 26 05:16:15.599: INFO: Error evaluating pod condition running and ready: want pod 'external-local-nodeport-k2k2v' on 'bootstrap-e2e-minion-group-vhdj' to be 'Running' but was 'Pending' Nov 26 05:16:17.616: INFO: Pod "external-local-nodeport-k2k2v": Phase="Pending", Reason="", readiness=false. Elapsed: 12.13475494s Nov 26 05:16:17.616: INFO: Error evaluating pod condition running and ready: want pod 'external-local-nodeport-k2k2v' on 'bootstrap-e2e-minion-group-vhdj' to be 'Running' but was 'Pending' Nov 26 05:16:19.623: INFO: Pod "external-local-nodeport-k2k2v": Phase="Pending", Reason="", readiness=false. Elapsed: 14.141315505s Nov 26 05:16:19.623: INFO: Error evaluating pod condition running and ready: want pod 'external-local-nodeport-k2k2v' on 'bootstrap-e2e-minion-group-vhdj' to be 'Running' but was 'Pending' Nov 26 05:16:21.603: INFO: Pod "external-local-nodeport-k2k2v": Phase="Pending", Reason="", readiness=false. Elapsed: 16.121458321s Nov 26 05:16:21.603: INFO: Error evaluating pod condition running and ready: want pod 'external-local-nodeport-k2k2v' on 'bootstrap-e2e-minion-group-vhdj' to be 'Running' but was 'Pending' Nov 26 05:16:23.605: INFO: Pod "external-local-nodeport-k2k2v": Phase="Running", Reason="", readiness=true. Elapsed: 18.123955277s Nov 26 05:16:23.605: INFO: Pod "external-local-nodeport-k2k2v" satisfied condition "running and ready" Nov 26 05:16:23.605: INFO: Wanted all 1 pods to be running and ready. Result: true. Pods: [external-local-nodeport-k2k2v] STEP: Performing setup for networking test in namespace esipp-5928 11/26/22 05:16:24.829 STEP: creating a selector 11/26/22 05:16:24.829 STEP: Creating the service pods in kubernetes 11/26/22 05:16:24.83 Nov 26 05:16:24.830: INFO: Waiting up to 10m0s for all (but 0) nodes to be schedulable Nov 26 05:16:25.200: INFO: Waiting up to 5m0s for pod "netserver-0" in namespace "esipp-5928" to be "running and ready" Nov 26 05:16:25.371: INFO: Pod "netserver-0": Phase="Pending", Reason="", readiness=false. Elapsed: 170.867102ms Nov 26 05:16:25.371: INFO: The phase of Pod netserver-0 is Pending, waiting for it to be Running (with Ready = true) Nov 26 05:16:27.453: INFO: Pod "netserver-0": Phase="Running", Reason="", readiness=false. Elapsed: 2.252383106s Nov 26 05:16:27.453: INFO: The phase of Pod netserver-0 is Running (Ready = false) Nov 26 05:16:29.435: INFO: Pod "netserver-0": Phase="Running", Reason="", readiness=false. Elapsed: 4.234531221s Nov 26 05:16:29.435: INFO: The phase of Pod netserver-0 is Running (Ready = false) Nov 26 05:16:31.423: INFO: Pod "netserver-0": Phase="Running", Reason="", readiness=false. Elapsed: 6.223243235s Nov 26 05:16:31.423: INFO: The phase of Pod netserver-0 is Running (Ready = false) Nov 26 05:16:33.426: INFO: Pod "netserver-0": Phase="Running", Reason="", readiness=false. Elapsed: 8.22547503s Nov 26 05:16:33.426: INFO: The phase of Pod netserver-0 is Running (Ready = false) Nov 26 05:16:35.482: INFO: Pod "netserver-0": Phase="Running", Reason="", readiness=false. Elapsed: 10.281849938s Nov 26 05:16:35.482: INFO: The phase of Pod netserver-0 is Running (Ready = false) Nov 26 05:16:37.427: INFO: Pod "netserver-0": Phase="Running", Reason="", readiness=false. Elapsed: 12.226927521s Nov 26 05:16:37.427: INFO: The phase of Pod netserver-0 is Running (Ready = false) Nov 26 05:16:39.447: INFO: Pod "netserver-0": Phase="Running", Reason="", readiness=false. Elapsed: 14.247273468s Nov 26 05:16:39.447: INFO: The phase of Pod netserver-0 is Running (Ready = false) Nov 26 05:16:41.420: INFO: Pod "netserver-0": Phase="Running", Reason="", readiness=false. Elapsed: 16.219553208s Nov 26 05:16:41.420: INFO: The phase of Pod netserver-0 is Running (Ready = false) Nov 26 05:16:43.422: INFO: Pod "netserver-0": Phase="Running", Reason="", readiness=false. Elapsed: 18.221548917s Nov 26 05:16:43.422: INFO: The phase of Pod netserver-0 is Running (Ready = false) Nov 26 05:16:45.428: INFO: Pod "netserver-0": Phase="Running", Reason="", readiness=false. Elapsed: 20.228364521s Nov 26 05:16:45.429: INFO: The phase of Pod netserver-0 is Running (Ready = false) Nov 26 05:16:47.432: INFO: Pod "netserver-0": Phase="Running", Reason="", readiness=true. Elapsed: 22.231900782s Nov 26 05:16:47.432: INFO: The phase of Pod netserver-0 is Running (Ready = true) Nov 26 05:16:47.432: INFO: Pod "netserver-0" satisfied condition "running and ready" Nov 26 05:16:47.481: INFO: Waiting up to 5m0s for pod "netserver-1" in namespace "esipp-5928" to be "running and ready" Nov 26 05:16:47.540: INFO: Pod "netserver-1": Phase="Running", Reason="", readiness=false. Elapsed: 58.225976ms Nov 26 05:16:47.540: INFO: The phase of Pod netserver-1 is Running (Ready = false) Nov 26 05:16:49.633: INFO: Pod "netserver-1": Phase="Running", Reason="", readiness=false. Elapsed: 2.151420741s Nov 26 05:16:49.633: INFO: The phase of Pod netserver-1 is Running (Ready = false) Nov 26 05:16:51.598: INFO: Pod "netserver-1": Phase="Running", Reason="", readiness=false. Elapsed: 4.117027112s Nov 26 05:16:51.598: INFO: The phase of Pod netserver-1 is Running (Ready = false) Nov 26 05:16:53.598: INFO: Pod "netserver-1": Phase="Running", Reason="", readiness=false. Elapsed: 6.116340643s Nov 26 05:16:53.598: INFO: The phase of Pod netserver-1 is Running (Ready = false) Nov 26 05:16:55.595: INFO: Pod "netserver-1": Phase="Running", Reason="", readiness=false. Elapsed: 8.113495783s Nov 26 05:16:55.595: INFO: The phase of Pod netserver-1 is Running (Ready = false) Nov 26 05:16:57.623: INFO: Pod "netserver-1": Phase="Running", Reason="", readiness=false. Elapsed: 10.141892763s Nov 26 05:16:57.623: INFO: The phase of Pod netserver-1 is Running (Ready = false) Nov 26 05:16:59.631: INFO: Pod "netserver-1": Phase="Running", Reason="", readiness=false. Elapsed: 12.149553619s Nov 26 05:16:59.631: INFO: The phase of Pod netserver-1 is Running (Ready = false) Nov 26 05:17:01.597: INFO: Pod "netserver-1": Phase="Running", Reason="", readiness=false. Elapsed: 14.115335205s Nov 26 05:17:01.597: INFO: The phase of Pod netserver-1 is Running (Ready = false) Nov 26 05:17:03.600: INFO: Pod "netserver-1": Phase="Running", Reason="", readiness=false. Elapsed: 16.119119187s Nov 26 05:17:03.600: INFO: The phase of Pod netserver-1 is Running (Ready = false) Nov 26 05:17:05.607: INFO: Pod "netserver-1": Phase="Running", Reason="", readiness=false. Elapsed: 18.12527109s Nov 26 05:17:05.607: INFO: The phase of Pod netserver-1 is Running (Ready = false) Nov 26 05:17:07.684: INFO: Pod "netserver-1": Phase="Running", Reason="", readiness=false. Elapsed: 20.203031625s Nov 26 05:17:07.684: INFO: The phase of Pod netserver-1 is Running (Ready = false) Nov 26 05:17:09.633: INFO: Pod "netserver-1": Phase="Running", Reason="", readiness=false. Elapsed: 22.152067369s Nov 26 05:17:09.633: INFO: The phase of Pod netserver-1 is Running (Ready = false) Nov 26 05:17:11.590: INFO: Pod "netserver-1": Phase="Running", Reason="", readiness=false. Elapsed: 24.108617976s Nov 26 05:17:11.590: INFO: The phase of Pod netserver-1 is Running (Ready = false) Nov 26 05:17:13.725: INFO: Pod "netserver-1": Phase="Running", Reason="", readiness=false. Elapsed: 26.243825852s Nov 26 05:17:13.725: INFO: The phase of Pod netserver-1 is Running (Ready = false) Nov 26 05:17:15.594: INFO: Pod "netserver-1": Phase="Running", Reason="", readiness=false. Elapsed: 28.112904488s Nov 26 05:17:15.594: INFO: The phase of Pod netserver-1 is Running (Ready = false) Nov 26 05:17:17.596: INFO: Pod "netserver-1": Phase="Running", Reason="", readiness=false. Elapsed: 30.114601185s Nov 26 05:17:17.596: INFO: The phase of Pod netserver-1 is Running (Ready = false) Nov 26 05:17:19.587: INFO: Pod "netserver-1": Phase="Running", Reason="", readiness=false. Elapsed: 32.105823688s Nov 26 05:17:19.587: INFO: The phase of Pod netserver-1 is Running (Ready = false) Nov 26 05:17:21.598: INFO: Pod "netserver-1": Phase="Running", Reason="", readiness=false. Elapsed: 34.116970456s Nov 26 05:17:21.598: INFO: The phase of Pod netserver-1 is Running (Ready = false) Nov 26 05:17:23.595: INFO: Pod "netserver-1": Phase="Running", Reason="", readiness=false. Elapsed: 36.113709243s Nov 26 05:17:23.595: INFO: The phase of Pod netserver-1 is Running (Ready = false) Nov 26 05:17:25.624: INFO: Pod "netserver-1": Phase="Running", Reason="", readiness=false. Elapsed: 38.142318157s Nov 26 05:17:25.624: INFO: The phase of Pod netserver-1 is Running (Ready = false) Nov 26 05:17:27.595: INFO: Pod "netserver-1": Phase="Running", Reason="", readiness=false. Elapsed: 40.113329522s Nov 26 05:17:27.595: INFO: The phase of Pod netserver-1 is Running (Ready = false) Nov 26 05:17:29.626: INFO: Pod "netserver-1": Phase="Running", Reason="", readiness=false. Elapsed: 42.144435653s Nov 26 05:17:29.626: INFO: The phase of Pod netserver-1 is Running (Ready = false) Nov 26 05:17:31.593: INFO: Pod "netserver-1": Phase="Running", Reason="", readiness=false. Elapsed: 44.112045823s Nov 26 05:17:31.593: INFO: The phase of Pod netserver-1 is Running (Ready = false) Nov 26 05:17:33.634: INFO: Pod "netserver-1": Phase="Running", Reason="", readiness=false. Elapsed: 46.152912157s Nov 26 05:17:33.634: INFO: The phase of Pod netserver-1 is Running (Ready = false) Nov 26 05:17:35.591: INFO: Pod "netserver-1": Phase="Running", Reason="", readiness=false. Elapsed: 48.110036252s Nov 26 05:17:35.591: INFO: The phase of Pod netserver-1 is Running (Ready = false) Nov 26 05:17:37.586: INFO: Pod "netserver-1": Phase="Running", Reason="", readiness=false. Elapsed: 50.104874442s Nov 26 05:17:37.586: INFO: The phase of Pod netserver-1 is Running (Ready = false) Nov 26 05:17:39.587: INFO: Pod "netserver-1": Phase="Running", Reason="", readiness=false. Elapsed: 52.105432371s Nov 26 05:17:39.587: INFO: The phase of Pod netserver-1 is Running (Ready = false) Nov 26 05:17:41.600: INFO: Pod "netserver-1": Phase="Running", Reason="", readiness=false. Elapsed: 54.118903014s Nov 26 05:17:41.600: INFO: The phase of Pod netserver-1 is Running (Ready = false) Nov 26 05:18:29.858: INFO: Pod "netserver-1": Phase="Running", Reason="", readiness=false. Elapsed: 1m42.377031456s Nov 26 05:18:29.858: INFO: The phase of Pod netserver-1 is Running (Ready = false) Nov 26 05:18:31.582: INFO: Pod "netserver-1": Phase="Running", Reason="", readiness=false. Elapsed: 1m44.100166169s Nov 26 05:18:31.582: INFO: The phase of Pod netserver-1 is Running (Ready = false) Nov 26 05:18:33.580: INFO: Pod "netserver-1": Phase="Running", Reason="", readiness=false. Elapsed: 1m46.098862594s Nov 26 05:18:33.580: INFO: The phase of Pod netserver-1 is Running (Ready = false) Nov 26 05:18:35.581: INFO: Pod "netserver-1": Phase="Running", Reason="", readiness=false. Elapsed: 1m48.099756879s Nov 26 05:18:35.581: INFO: The phase of Pod netserver-1 is Running (Ready = false) Nov 26 05:18:37.580: INFO: Pod "netserver-1": Phase="Running", Reason="", readiness=false. Elapsed: 1m50.099028572s Nov 26 05:18:37.580: INFO: The phase of Pod netserver-1 is Running (Ready = false) Nov 26 05:18:39.581: INFO: Pod "netserver-1": Phase="Running", Reason="", readiness=false. Elapsed: 1m52.0993623s Nov 26 05:18:39.581: INFO: The phase of Pod netserver-1 is Running (Ready = false) Nov 26 05:18:41.581: INFO: Pod "netserver-1": Phase="Running", Reason="", readiness=false. Elapsed: 1m54.099308926s Nov 26 05:18:41.581: INFO: The phase of Pod netserver-1 is Running (Ready = false) Nov 26 05:18:43.581: INFO: Pod "netserver-1": Phase="Running", Reason="", readiness=false. Elapsed: 1m56.099833722s Nov 26 05:18:43.581: INFO: The phase of Pod netserver-1 is Running (Ready = false) Nov 26 05:18:45.581: INFO: Pod "netserver-1": Phase="Running", Reason="", readiness=false. Elapsed: 1m58.099251927s Nov 26 05:18:45.581: INFO: The phase of Pod netserver-1 is Running (Ready = false) Nov 26 05:18:47.581: INFO: Pod "netserver-1": Phase="Running", Reason="", readiness=false. Elapsed: 2m0.09915682s Nov 26 05:18:47.581: INFO: The phase of Pod netserver-1 is Running (Ready = false) Nov 26 05:18:49.580: INFO: Pod "netserver-1": Phase="Running", Reason="", readiness=false. Elapsed: 2m2.099112688s Nov 26 05:18:49.581: INFO: The phase of Pod netserver-1 is Running (Ready = false) Nov 26 05:18:51.581: INFO: Pod "netserver-1": Phase="Running", Reason="", readiness=false. Elapsed: 2m4.100049842s Nov 26 05:18:51.581: INFO: The phase of Pod netserver-1 is Running (Ready = false) Nov 26 05:18:53.581: INFO: Pod "netserver-1": Phase="Running", Reason="", readiness=false. Elapsed: 2m6.099551083s Nov 26 05:18:53.581: INFO: The phase of Pod netserver-1 is Running (Ready = false) Nov 26 05:18:55.581: INFO: Pod "netserver-1": Phase="Running", Reason="", readiness=false. Elapsed: 2m8.099177587s Nov 26 05:18:55.581: INFO: The phase of Pod netserver-1 is Running (Ready = false) Nov 26 05:18:57.580: INFO: Pod "netserver-1": Phase="Running", Reason="", readiness=false. Elapsed: 2m10.099140008s Nov 26 05:18:57.581: INFO: The phase of Pod netserver-1 is Running (Ready = false) Nov 26 05:18:59.581: INFO: Pod "netserver-1": Phase="Running", Reason="", readiness=false. Elapsed: 2m12.09915329s Nov 26 05:18:59.581: INFO: The phase of Pod netserver-1 is Running (Ready = false) Nov 26 05:19:01.583: INFO: Pod "netserver-1": Phase="Running", Reason="", readiness=false. Elapsed: 2m14.10165586s Nov 26 05:19:01.583: INFO: The phase of Pod netserver-1 is Running (Ready = false) Nov 26 05:19:03.593: INFO: Pod "netserver-1": Phase="Running", Reason="", readiness=false. Elapsed: 2m16.111919417s Nov 26 05:19:03.593: INFO: The phase of Pod netserver-1 is Running (Ready = false) Nov 26 05:19:05.580: INFO: Pod "netserver-1": Phase="Running", Reason="", readiness=false. Elapsed: 2m18.098806404s Nov 26 05:19:05.580: INFO: The phase of Pod netserver-1 is Running (Ready = false) Nov 26 05:19:07.581: INFO: Pod "netserver-1": Phase="Running", Reason="", readiness=false. Elapsed: 2m20.100004684s Nov 26 05:19:07.581: INFO: The phase of Pod netserver-1 is Running (Ready = false) Nov 26 05:19:09.581: INFO: Pod "netserver-1": Phase="Running", Reason="", readiness=false. Elapsed: 2m22.099640818s Nov 26 05:19:09.581: INFO: The phase of Pod netserver-1 is Running (Ready = false) Nov 26 05:19:11.583: INFO: Pod "netserver-1": Phase="Running", Reason="", readiness=false. Elapsed: 2m24.101325221s Nov 26 05:19:11.583: INFO: The phase of Pod netserver-1 is Running (Ready = false) Nov 26 05:19:13.580: INFO: Pod "netserver-1": Phase="Running", Reason="", readiness=false. Elapsed: 2m26.098671296s Nov 26 05:19:13.580: INFO: The phase of Pod netserver-1 is Running (Ready = false) Nov 26 05:19:15.607: INFO: Pod "netserver-1": Phase="Running", Reason="", readiness=false. Elapsed: 2m28.1254488s Nov 26 05:19:15.607: INFO: The phase of Pod netserver-1 is Running (Ready = false) Nov 26 05:19:17.630: INFO: Pod "netserver-1": Phase="Running", Reason="", readiness=false. Elapsed: 2m30.149014511s Nov 26 05:19:17.630: INFO: The phase of Pod netserver-1 is Running (Ready = false) Nov 26 05:19:19.580: INFO: Pod "netserver-1": Phase="Running", Reason="", readiness=false. Elapsed: 2m32.099106709s Nov 26 05:19:19.580: INFO: The phase of Pod netserver-1 is Running (Ready = false) Nov 26 05:19:21.582: INFO: Pod "netserver-1": Phase="Running", Reason="", readiness=false. Elapsed: 2m34.10076784s Nov 26 05:19:21.582: INFO: The phase of Pod netserver-1 is Running (Ready = false) Nov 26 05:19:23.580: INFO: Pod "netserver-1": Phase="Running", Reason="", readiness=false. Elapsed: 2m36.098815992s Nov 26 05:19:23.580: INFO: The phase of Pod netserver-1 is Running (Ready = false) Nov 26 05:19:25.580: INFO: Pod "netserver-1": Phase="Running", Reason="", readiness=false. Elapsed: 2m38.098988293s Nov 26 05:19:25.580: INFO: The phase of Pod netserver-1 is Running (Ready = false) Nov 26 05:19:27.581: INFO: Pod "netserver-1": Phase="Running", Reason="", readiness=false. Elapsed: 2m40.099720518s Nov 26 05:19:27.581: INFO: The phase of Pod netserver-1 is Running (Ready = false) Nov 26 05:19:29.580: INFO: Pod "netserver-1": Phase="Running", Reason="", readiness=false. Elapsed: 2m42.098990508s Nov 26 05:19:29.580: INFO: The phase of Pod netserver-1 is Running (Ready = false) Nov 26 05:19:31.588: INFO: Pod "netserver-1": Phase="Running", Reason="", readiness=false. Elapsed: 2m44.10705334s Nov 26 05:19:31.588: INFO: The phase of Pod netserver-1 is Running (Ready = false) Nov 26 05:19:33.592: INFO: Pod "netserver-1": Phase="Running", Reason="", readiness=false. Elapsed: 2m46.110241957s Nov 26 05:19:33.592: INFO: The phase of Pod netserver-1 is Running (Ready = false) Nov 26 05:19:35.580: INFO: Pod "netserver-1": Phase="Running", Reason="", readiness=false. Elapsed: 2m48.098980547s Nov 26 05:19:35.580: INFO: The phase of Pod netserver-1 is Running (Ready = false) Nov 26 05:19:37.581: INFO: Pod "netserver-1": Phase="Running", Reason="", readiness=false. Elapsed: 2m50.099986789s Nov 26 05:19:37.581: INFO: The phase of Pod netserver-1 is Running (Ready = false) Nov 26 05:19:39.581: INFO: Pod "netserver-1": Phase="Running", Reason="", readiness=false. Elapsed: 2m52.100094843s Nov 26 05:19:39.581: INFO: The phase of Pod netserver-1 is Running (Ready = false) Nov 26 05:20:41.583: INFO: Encountered non-retryable error while getting pod esipp-5928/netserver-1: Get "https://35.227.179.144/api/v1/namespaces/esipp-5928/pods/netserver-1": stream error: stream ID 241; INTERNAL_ERROR; received from peer Nov 26 05:20:41.583: INFO: Unexpected error: <*fmt.wrapError | 0xc000124860>: { msg: "error while waiting for pod esipp-5928/netserver-1 to be running and ready: Get \"https://35.227.179.144/api/v1/namespaces/esipp-5928/pods/netserver-1\": stream error: stream ID 241; INTERNAL_ERROR; received from peer", err: <*url.Error | 0xc0037a1200>{ Op: "Get", URL: "https://35.227.179.144/api/v1/namespaces/esipp-5928/pods/netserver-1", Err: <http2.StreamError>{ StreamID: 241, Code: 2, Cause: <*errors.errorString | 0xc00017d550>{ s: "received from peer", }, }, }, } Nov 26 05:20:41.583: FAIL: error while waiting for pod esipp-5928/netserver-1 to be running and ready: Get "https://35.227.179.144/api/v1/namespaces/esipp-5928/pods/netserver-1": stream error: stream ID 241; INTERNAL_ERROR; received from peer Full Stack Trace k8s.io/kubernetes/test/e2e/framework/network.(*NetworkingTestConfig).createNetProxyPods(0xc0010147e0, {0x75c6f7c, 0x9}, 0xc0022e0d80) test/e2e/framework/network/utils.go:866 +0x1d0 k8s.io/kubernetes/test/e2e/framework/network.(*NetworkingTestConfig).setupCore(0xc0010147e0, 0x7f18e001c5e0?) test/e2e/framework/network/utils.go:763 +0x55 k8s.io/kubernetes/test/e2e/framework/network.(*NetworkingTestConfig).setup(0xc0010147e0, 0x3c?) test/e2e/framework/network/utils.go:778 +0x3e k8s.io/kubernetes/test/e2e/framework/network.NewNetworkingTestConfig(0xc000cc0000, {0x0, 0x0, 0x0?}) test/e2e/framework/network/utils.go:131 +0x125 k8s.io/kubernetes/test/e2e/network.glob..func20.4() test/e2e/network/loadbalancer.go:1332 +0x145 [AfterEach] [sig-network] LoadBalancers ESIPP [Slow] test/e2e/framework/node/init/init.go:32 Nov 26 05:20:49.477: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready [AfterEach] [sig-network] LoadBalancers ESIPP [Slow] test/e2e/network/loadbalancer.go:1260 Nov 26 05:20:49.526: INFO: Output of kubectl describe svc: Nov 26 05:20:49.526: INFO: Running '/workspace/github.com/containerd/containerd/kubernetes/platforms/linux/amd64/kubectl --server=https://35.227.179.144 --kubeconfig=/workspace/.kube/config --namespace=esipp-5928 describe svc --namespace=esipp-5928' Nov 26 05:20:49.734: INFO: stderr: "No resources found in esipp-5928 namespace.\n" Nov 26 05:20:49.734: INFO: stdout: "" Nov 26 05:20:49.734: INFO: [DeferCleanup (Each)] [sig-network] LoadBalancers ESIPP [Slow] test/e2e/framework/metrics/init/init.go:33 [DeferCleanup (Each)] [sig-network] LoadBalancers ESIPP [Slow] dump namespaces | framework.go:196 STEP: dump namespace information after failure 11/26/22 05:20:49.734 STEP: Collecting events from namespace "esipp-5928". 11/26/22 05:20:49.734 STEP: Found 29 events. 11/26/22 05:20:49.779 Nov 26 05:20:49.779: INFO: At 0001-01-01 00:00:00 +0000 UTC - event for external-local-nodeport-k2k2v: { } Scheduled: Successfully assigned esipp-5928/external-local-nodeport-k2k2v to bootstrap-e2e-minion-group-vhdj Nov 26 05:20:49.779: INFO: At 0001-01-01 00:00:00 +0000 UTC - event for netserver-0: { } Scheduled: Successfully assigned esipp-5928/netserver-0 to bootstrap-e2e-minion-group-0975 Nov 26 05:20:49.779: INFO: At 0001-01-01 00:00:00 +0000 UTC - event for netserver-1: { } Scheduled: Successfully assigned esipp-5928/netserver-1 to bootstrap-e2e-minion-group-rwsn Nov 26 05:20:49.779: INFO: At 0001-01-01 00:00:00 +0000 UTC - event for netserver-2: { } Scheduled: Successfully assigned esipp-5928/netserver-2 to bootstrap-e2e-minion-group-vhdj Nov 26 05:20:49.779: INFO: At 2022-11-26 05:16:03 +0000 UTC - event for external-local-nodeport: {replication-controller } SuccessfulCreate: Created pod: external-local-nodeport-k2k2v Nov 26 05:20:49.779: INFO: At 2022-11-26 05:16:05 +0000 UTC - event for external-local-nodeport-k2k2v: {kubelet bootstrap-e2e-minion-group-vhdj} Pulled: Container image "registry.k8s.io/e2e-test-images/agnhost:2.43" already present on machine Nov 26 05:20:49.779: INFO: At 2022-11-26 05:16:05 +0000 UTC - event for external-local-nodeport-k2k2v: {kubelet bootstrap-e2e-minion-group-vhdj} Created: Created container netexec Nov 26 05:20:49.779: INFO: At 2022-11-26 05:16:06 +0000 UTC - event for external-local-nodeport-k2k2v: {kubelet bootstrap-e2e-minion-group-vhdj} Started: Started container netexec Nov 26 05:20:49.779: INFO: At 2022-11-26 05:16:08 +0000 UTC - event for external-local-nodeport-k2k2v: {kubelet bootstrap-e2e-minion-group-vhdj} Killing: Stopping container netexec Nov 26 05:20:49.779: INFO: At 2022-11-26 05:16:10 +0000 UTC - event for external-local-nodeport-k2k2v: {kubelet bootstrap-e2e-minion-group-vhdj} Unhealthy: Readiness probe failed: Get "http://10.64.3.105:80/hostName": dial tcp 10.64.3.105:80: connect: connection refused Nov 26 05:20:49.779: INFO: At 2022-11-26 05:16:11 +0000 UTC - event for external-local-nodeport-k2k2v: {kubelet bootstrap-e2e-minion-group-vhdj} SandboxChanged: Pod sandbox changed, it will be killed and re-created. Nov 26 05:20:49.779: INFO: At 2022-11-26 05:16:26 +0000 UTC - event for netserver-0: {kubelet bootstrap-e2e-minion-group-0975} Started: Started container webserver Nov 26 05:20:49.779: INFO: At 2022-11-26 05:16:26 +0000 UTC - event for netserver-0: {kubelet bootstrap-e2e-minion-group-0975} Killing: Stopping container webserver Nov 26 05:20:49.779: INFO: At 2022-11-26 05:16:26 +0000 UTC - event for netserver-0: {kubelet bootstrap-e2e-minion-group-0975} Created: Created container webserver Nov 26 05:20:49.779: INFO: At 2022-11-26 05:16:26 +0000 UTC - event for netserver-0: {kubelet bootstrap-e2e-minion-group-0975} Pulled: Container image "registry.k8s.io/e2e-test-images/agnhost:2.43" already present on machine Nov 26 05:20:49.779: INFO: At 2022-11-26 05:16:26 +0000 UTC - event for netserver-2: {kubelet bootstrap-e2e-minion-group-vhdj} Created: Created container webserver Nov 26 05:20:49.779: INFO: At 2022-11-26 05:16:26 +0000 UTC - event for netserver-2: {kubelet bootstrap-e2e-minion-group-vhdj} Started: Started container webserver Nov 26 05:20:49.779: INFO: At 2022-11-26 05:16:26 +0000 UTC - event for netserver-2: {kubelet bootstrap-e2e-minion-group-vhdj} Pulled: Container image "registry.k8s.io/e2e-test-images/agnhost:2.43" already present on machine Nov 26 05:20:49.779: INFO: At 2022-11-26 05:16:27 +0000 UTC - event for netserver-0: {kubelet bootstrap-e2e-minion-group-0975} SandboxChanged: Pod sandbox changed, it will be killed and re-created. Nov 26 05:20:49.779: INFO: At 2022-11-26 05:16:27 +0000 UTC - event for netserver-1: {kubelet bootstrap-e2e-minion-group-rwsn} Pulled: Container image "registry.k8s.io/e2e-test-images/agnhost:2.43" already present on machine Nov 26 05:20:49.779: INFO: At 2022-11-26 05:16:27 +0000 UTC - event for netserver-1: {kubelet bootstrap-e2e-minion-group-rwsn} Created: Created container webserver Nov 26 05:20:49.779: INFO: At 2022-11-26 05:16:27 +0000 UTC - event for netserver-1: {kubelet bootstrap-e2e-minion-group-rwsn} Started: Started container webserver Nov 26 05:20:49.779: INFO: At 2022-11-26 05:16:29 +0000 UTC - event for netserver-1: {kubelet bootstrap-e2e-minion-group-rwsn} Killing: Stopping container webserver Nov 26 05:20:49.779: INFO: At 2022-11-26 05:16:30 +0000 UTC - event for netserver-1: {kubelet bootstrap-e2e-minion-group-rwsn} SandboxChanged: Pod sandbox changed, it will be killed and re-created. Nov 26 05:20:49.779: INFO: At 2022-11-26 05:16:33 +0000 UTC - event for netserver-1: {kubelet bootstrap-e2e-minion-group-rwsn} BackOff: Back-off restarting failed container webserver in pod netserver-1_esipp-5928(fa0494ae-494a-4b5c-87a4-3d6a48859884) Nov 26 05:20:49.779: INFO: At 2022-11-26 05:17:24 +0000 UTC - event for external-local-nodeport-k2k2v: {kubelet bootstrap-e2e-minion-group-vhdj} BackOff: Back-off restarting failed container netexec in pod external-local-nodeport-k2k2v_esipp-5928(d1e0c5bf-1f1d-4f1e-9cbf-02c885195092) Nov 26 05:20:49.779: INFO: At 2022-11-26 05:17:51 +0000 UTC - event for netserver-2: {kubelet bootstrap-e2e-minion-group-vhdj} Killing: Stopping container webserver Nov 26 05:20:49.779: INFO: At 2022-11-26 05:17:52 +0000 UTC - event for netserver-2: {kubelet bootstrap-e2e-minion-group-vhdj} SandboxChanged: Pod sandbox changed, it will be killed and re-created. Nov 26 05:20:49.779: INFO: At 2022-11-26 05:17:54 +0000 UTC - event for netserver-2: {kubelet bootstrap-e2e-minion-group-vhdj} BackOff: Back-off restarting failed container webserver in pod netserver-2_esipp-5928(5478087a-9205-4e58-9101-2c36c6489bc4) Nov 26 05:20:56.046: INFO: POD NODE PHASE GRACE CONDITIONS Nov 26 05:20:56.046: INFO: external-local-nodeport-k2k2v bootstrap-e2e-minion-group-vhdj Running [{Initialized True 0001-01-01 00:00:00 +0000 UTC 2022-11-26 05:16:03 +0000 UTC } {Ready True 0001-01-01 00:00:00 +0000 UTC 2022-11-26 05:17:38 +0000 UTC } {ContainersReady True 0001-01-01 00:00:00 +0000 UTC 2022-11-26 05:17:38 +0000 UTC } {PodScheduled True 0001-01-01 00:00:00 +0000 UTC 2022-11-26 05:16:03 +0000 UTC }] Nov 26 05:20:56.046: INFO: netserver-0 bootstrap-e2e-minion-group-0975 Running [{Initialized True 0001-01-01 00:00:00 +0000 UTC 2022-11-26 05:16:25 +0000 UTC } {Ready True 0001-01-01 00:00:00 +0000 UTC 2022-11-26 05:16:45 +0000 UTC } {ContainersReady True 0001-01-01 00:00:00 +0000 UTC 2022-11-26 05:16:45 +0000 UTC } {PodScheduled True 0001-01-01 00:00:00 +0000 UTC 2022-11-26 05:16:25 +0000 UTC }] Nov 26 05:20:56.046: INFO: netserver-1 bootstrap-e2e-minion-group-rwsn Running [{Initialized True 0001-01-01 00:00:00 +0000 UTC 2022-11-26 05:16:25 +0000 UTC } {Ready False 0001-01-01 00:00:00 +0000 UTC 2022-11-26 05:17:06 +0000 UTC ContainersNotReady containers with unready status: [webserver]} {ContainersReady False 0001-01-01 00:00:00 +0000 UTC 2022-11-26 05:17:06 +0000 UTC ContainersNotReady containers with unready status: [webserver]} {PodScheduled True 0001-01-01 00:00:00 +0000 UTC 2022-11-26 05:16:25 +0000 UTC }] Nov 26 05:20:56.046: INFO: netserver-2 bootstrap-e2e-minion-group-vhdj Running [{Initialized True 0001-01-01 00:00:00 +0000 UTC 2022-11-26 05:16:25 +0000 UTC } {Ready False 0001-01-01 00:00:00 +0000 UTC 2022-11-26 05:19:26 +0000 UTC ContainersNotReady containers with unready status: [webserver]} {ContainersReady False 0001-01-01 00:00:00 +0000 UTC 2022-11-26 05:19:26 +0000 UTC ContainersNotReady containers with unready status: [webserver]} {PodScheduled True 0001-01-01 00:00:00 +0000 UTC 2022-11-26 05:16:25 +0000 UTC }] Nov 26 05:20:56.046: INFO: Nov 26 05:20:56.849: INFO: Logging node info for node bootstrap-e2e-master Nov 26 05:20:56.913: INFO: Node Info: &Node{ObjectMeta:{bootstrap-e2e-master 9a48eb11-c964-4bf8-82c0-1c32a9703981 7221 0 2022-11-26 05:05:18 +0000 UTC <nil> <nil> map[beta.kubernetes.io/arch:amd64 beta.kubernetes.io/instance-type:n1-standard-1 beta.kubernetes.io/os:linux cloud.google.com/metadata-proxy-ready:true failure-domain.beta.kubernetes.io/region:us-west1 failure-domain.beta.kubernetes.io/zone:us-west1-b kubernetes.io/arch:amd64 kubernetes.io/hostname:bootstrap-e2e-master kubernetes.io/os:linux node.kubernetes.io/instance-type:n1-standard-1 topology.kubernetes.io/region:us-west1 topology.kubernetes.io/zone:us-west1-b] map[node.alpha.kubernetes.io/ttl:0 volumes.kubernetes.io/controller-managed-attach-detach:true] [] [] [{kubelet Update v1 2022-11-26 05:05:18 +0000 UTC FieldsV1 {"f:metadata":{"f:annotations":{".":{},"f:volumes.kubernetes.io/controller-managed-attach-detach":{}},"f:labels":{".":{},"f:beta.kubernetes.io/arch":{},"f:beta.kubernetes.io/instance-type":{},"f:beta.kubernetes.io/os":{},"f:cloud.google.com/metadata-proxy-ready":{},"f:failure-domain.beta.kubernetes.io/region":{},"f:failure-domain.beta.kubernetes.io/zone":{},"f:kubernetes.io/arch":{},"f:kubernetes.io/hostname":{},"f:kubernetes.io/os":{},"f:node.kubernetes.io/instance-type":{},"f:topology.kubernetes.io/region":{},"f:topology.kubernetes.io/zone":{}}},"f:spec":{"f:providerID":{},"f:unschedulable":{}}} } {kube-controller-manager Update v1 2022-11-26 05:05:25 +0000 UTC FieldsV1 {"f:status":{"f:conditions":{"k:{\"type\":\"NetworkUnavailable\"}":{"f:lastHeartbeatTime":{},"f:lastTransitionTime":{},"f:message":{},"f:reason":{},"f:status":{}}}}} status} {kube-controller-manager Update v1 2022-11-26 05:05:39 +0000 UTC FieldsV1 {"f:metadata":{"f:annotations":{"f:node.alpha.kubernetes.io/ttl":{}}},"f:spec":{"f:podCIDR":{},"f:podCIDRs":{".":{},"v:\"10.64.1.0/24\"":{}},"f:taints":{}}} } {kubelet Update v1 2022-11-26 05:16:15 +0000 UTC FieldsV1 {"f:status":{"f:conditions":{"k:{\"type\":\"DiskPressure\"}":{"f:lastHeartbeatTime":{}},"k:{\"type\":\"MemoryPressure\"}":{"f:lastHeartbeatTime":{}},"k:{\"type\":\"PIDPressure\"}":{"f:lastHeartbeatTime":{}},"k:{\"type\":\"Ready\"}":{"f:lastHeartbeatTime":{},"f:lastTransitionTime":{},"f:message":{},"f:reason":{},"f:status":{}}},"f:images":{}}} status}]},Spec:NodeSpec{PodCIDR:10.64.1.0/24,DoNotUseExternalID:,ProviderID:gce://k8s-jkns-e2e-gce-ci-reboot/us-west1-b/bootstrap-e2e-master,Unschedulable:true,Taints:[]Taint{Taint{Key:node-role.kubernetes.io/master,Value:,Effect:NoSchedule,TimeAdded:<nil>,},Taint{Key:node.kubernetes.io/unschedulable,Value:,Effect:NoSchedule,TimeAdded:<nil>,},},ConfigSource:nil,PodCIDRs:[10.64.1.0/24],},Status:NodeStatus{Capacity:ResourceList{attachable-volumes-gce-pd: {{127 0} {<nil>} 127 DecimalSI},cpu: {{1 0} {<nil>} 1 DecimalSI},ephemeral-storage: {{16656896000 0} {<nil>} 16266500Ki BinarySI},hugepages-1Gi: {{0 0} {<nil>} 0 DecimalSI},hugepages-2Mi: {{0 0} {<nil>} 0 DecimalSI},memory: {{3858374656 0} {<nil>} BinarySI},pods: {{110 0} {<nil>} 110 DecimalSI},},Allocatable:ResourceList{attachable-volumes-gce-pd: {{127 0} {<nil>} 127 DecimalSI},cpu: {{1 0} {<nil>} 1 DecimalSI},ephemeral-storage: {{14991206376 0} {<nil>} 14991206376 DecimalSI},hugepages-1Gi: {{0 0} {<nil>} 0 DecimalSI},hugepages-2Mi: {{0 0} {<nil>} 0 DecimalSI},memory: {{3596230656 0} {<nil>} BinarySI},pods: {{110 0} {<nil>} 110 DecimalSI},},Phase:,Conditions:[]NodeCondition{NodeCondition{Type:NetworkUnavailable,Status:False,LastHeartbeatTime:2022-11-26 05:05:25 +0000 UTC,LastTransitionTime:2022-11-26 05:05:25 +0000 UTC,Reason:RouteCreated,Message:RouteController created a route,},NodeCondition{Type:MemoryPressure,Status:False,LastHeartbeatTime:2022-11-26 05:16:15 +0000 UTC,LastTransitionTime:2022-11-26 05:05:18 +0000 UTC,Reason:KubeletHasSufficientMemory,Message:kubelet has sufficient memory available,},NodeCondition{Type:DiskPressure,Status:False,LastHeartbeatTime:2022-11-26 05:16:15 +0000 UTC,LastTransitionTime:2022-11-26 05:05:18 +0000 UTC,Reason:KubeletHasNoDiskPressure,Message:kubelet has no disk pressure,},NodeCondition{Type:PIDPressure,Status:False,LastHeartbeatTime:2022-11-26 05:16:15 +0000 UTC,LastTransitionTime:2022-11-26 05:05:18 +0000 UTC,Reason:KubeletHasSufficientPID,Message:kubelet has sufficient PID available,},NodeCondition{Type:Ready,Status:True,LastHeartbeatTime:2022-11-26 05:16:15 +0000 UTC,LastTransitionTime:2022-11-26 05:05:38 +0000 UTC,Reason:KubeletReady,Message:kubelet is posting ready status. AppArmor enabled,},},Addresses:[]NodeAddress{NodeAddress{Type:InternalIP,Address:10.138.0.2,},NodeAddress{Type:ExternalIP,Address:35.227.179.144,},NodeAddress{Type:InternalDNS,Address:bootstrap-e2e-master.c.k8s-jkns-e2e-gce-ci-reboot.internal,},NodeAddress{Type:Hostname,Address:bootstrap-e2e-master.c.k8s-jkns-e2e-gce-ci-reboot.internal,},},DaemonEndpoints:NodeDaemonEndpoints{KubeletEndpoint:DaemonEndpoint{Port:10250,},},NodeInfo:NodeSystemInfo{MachineID:0fdb3cfe29f66637553465718381a2f8,SystemUUID:0fdb3cfe-29f6-6637-5534-65718381a2f8,BootID:fb804a29-7e61-434d-8044-6944159ce35f,KernelVersion:5.10.123+,OSImage:Container-Optimized OS from Google,ContainerRuntimeVersion:containerd://1.7.0-beta.0-149-gd06318622,KubeletVersion:v1.27.0-alpha.0.50+70617042976dc1,KubeProxyVersion:v1.27.0-alpha.0.50+70617042976dc1,OperatingSystem:linux,Architecture:amd64,},Images:[]ContainerImage{ContainerImage{Names:[registry.k8s.io/kube-apiserver-amd64:v1.27.0-alpha.0.50_70617042976dc1],SizeBytes:135160272,},ContainerImage{Names:[registry.k8s.io/kube-controller-manager-amd64:v1.27.0-alpha.0.50_70617042976dc1],SizeBytes:124990265,},ContainerImage{Names:[registry.k8s.io/etcd@sha256:dd75ec974b0a2a6f6bb47001ba09207976e625db898d1b16735528c009cb171c registry.k8s.io/etcd:3.5.6-0],SizeBytes:102542580,},ContainerImage{Names:[registry.k8s.io/kube-scheduler-amd64:v1.27.0-alpha.0.50_70617042976dc1],SizeBytes:57660216,},ContainerImage{Names:[gke.gcr.io/prometheus-to-sd@sha256:e739643c3939ba0b161425f45a1989eedfc4a3b166db9a7100863296b4c70510 gke.gcr.io/prometheus-to-sd:v0.11.1-gke.1],SizeBytes:48742566,},ContainerImage{Names:[gcr.io/k8s-ingress-image-push/ingress-gce-glbc-amd64@sha256:5db27383add6d9f4ebdf0286409ac31f7f5d273690204b341a4e37998917693b gcr.io/k8s-ingress-image-push/ingress-gce-glbc-amd64:v1.20.1],SizeBytes:36598135,},ContainerImage{Names:[registry.k8s.io/addon-manager/kube-addon-manager@sha256:49cc4e6e4a3745b427ce14b0141476ab339bb65c6bc05033019e046c8727dcb0 registry.k8s.io/addon-manager/kube-addon-manager:v9.1.6],SizeBytes:30464183,},ContainerImage{Names:[registry.k8s.io/kas-network-proxy/proxy-server@sha256:2c111f004bec24888d8cfa2a812a38fb8341350abac67dcd0ac64e709dfe389c registry.k8s.io/kas-network-proxy/proxy-server:v0.0.33],SizeBytes:22020129,},ContainerImage{Names:[registry.k8s.io/metadata-proxy@sha256:e914645f22e946bce5165737e1b244e0a296ad1f0f81a9531adc57af2780978a registry.k8s.io/metadata-proxy:v0.1.12],SizeBytes:5301657,},ContainerImage{Names:[registry.k8s.io/pause@sha256:9001185023633d17a2f98ff69b6ff2615b8ea02a825adffa40422f51dfdcde9d registry.k8s.io/pause:3.8],SizeBytes:311286,},},VolumesInUse:[],VolumesAttached:[]AttachedVolume{},Config:nil,},} Nov 26 05:20:56.914: INFO: Logging kubelet events for node bootstrap-e2e-master Nov 26 05:20:57.050: INFO: Logging pods the kubelet thinks is on node bootstrap-e2e-master Nov 26 05:20:57.149: INFO: etcd-server-events-bootstrap-e2e-master started at 2022-11-26 05:04:35 +0000 UTC (0+1 container statuses recorded) Nov 26 05:20:57.149: INFO: Container etcd-container ready: true, restart count 3 Nov 26 05:20:57.149: INFO: kube-controller-manager-bootstrap-e2e-master started at 2022-11-26 05:04:35 +0000 UTC (0+1 container statuses recorded) Nov 26 05:20:57.149: INFO: Container kube-controller-manager ready: false, restart count 5 Nov 26 05:20:57.149: INFO: kube-scheduler-bootstrap-e2e-master started at 2022-11-26 05:04:35 +0000 UTC (0+1 container statuses recorded) Nov 26 05:20:57.149: INFO: Container kube-scheduler ready: false, restart count 5 Nov 26 05:20:57.149: INFO: kube-addon-manager-bootstrap-e2e-master started at 2022-11-26 05:04:52 +0000 UTC (0+1 container statuses recorded) Nov 26 05:20:57.149: INFO: Container kube-addon-manager ready: true, restart count 1 Nov 26 05:20:57.149: INFO: metadata-proxy-v0.1-8ll2g started at 2022-11-26 05:05:51 +0000 UTC (0+2 container statuses recorded) Nov 26 05:20:57.149: INFO: Container metadata-proxy ready: true, restart count 0 Nov 26 05:20:57.149: INFO: Container prometheus-to-sd-exporter ready: true, restart count 0 Nov 26 05:20:57.149: INFO: etcd-server-bootstrap-e2e-master started at 2022-11-26 05:04:35 +0000 UTC (0+1 container statuses recorded) Nov 26 05:20:57.149: INFO: Container etcd-container ready: true, restart count 4 Nov 26 05:20:57.149: INFO: konnectivity-server-bootstrap-e2e-master started at 2022-11-26 05:04:35 +0000 UTC (0+1 container statuses recorded) Nov 26 05:20:57.149: INFO: Container konnectivity-server-container ready: true, restart count 2 Nov 26 05:20:57.149: INFO: kube-apiserver-bootstrap-e2e-master started at 2022-11-26 05:04:35 +0000 UTC (0+1 container statuses recorded) Nov 26 05:20:57.149: INFO: Container kube-apiserver ready: true, restart count 1 Nov 26 05:20:57.149: INFO: l7-lb-controller-bootstrap-e2e-master started at 2022-11-26 05:04:52 +0000 UTC (0+1 container statuses recorded) Nov 26 05:20:57.149: INFO: Container l7-lb-controller ready: false, restart count 6 Nov 26 05:20:57.413: INFO: Latency metrics for node bootstrap-e2e-master Nov 26 05:20:57.413: INFO: Logging node info for node bootstrap-e2e-minion-group-0975 Nov 26 05:20:57.465: INFO: Node Info: &Node{ObjectMeta:{bootstrap-e2e-minion-group-0975 6a0671d4-0e9d-4ac8-b359-65111262cfad 8761 0 2022-11-26 05:05:22 +0000 UTC <nil> <nil> map[beta.kubernetes.io/arch:amd64 beta.kubernetes.io/instance-type:n1-standard-2 beta.kubernetes.io/os:linux cloud.google.com/metadata-proxy-ready:true failure-domain.beta.kubernetes.io/region:us-west1 failure-domain.beta.kubernetes.io/zone:us-west1-b kubernetes.io/arch:amd64 kubernetes.io/hostname:bootstrap-e2e-minion-group-0975 kubernetes.io/os:linux node.kubernetes.io/instance-type:n1-standard-2 topology.hostpath.csi/node:bootstrap-e2e-minion-group-0975 topology.kubernetes.io/region:us-west1 topology.kubernetes.io/zone:us-west1-b] map[csi.volume.kubernetes.io/nodeid:{"csi-hostpath-multivolume-5830":"bootstrap-e2e-minion-group-0975","csi-hostpath-multivolume-593":"bootstrap-e2e-minion-group-0975","csi-hostpath-multivolume-702":"bootstrap-e2e-minion-group-0975","csi-hostpath-provisioning-7492":"bootstrap-e2e-minion-group-0975","csi-mock-csi-mock-volumes-8067":"csi-mock-csi-mock-volumes-8067"} node.alpha.kubernetes.io/ttl:0 volumes.kubernetes.io/controller-managed-attach-detach:true] [] [] [{kubelet Update v1 2022-11-26 05:05:22 +0000 UTC FieldsV1 {"f:metadata":{"f:annotations":{".":{},"f:volumes.kubernetes.io/controller-managed-attach-detach":{}},"f:labels":{".":{},"f:beta.kubernetes.io/arch":{},"f:beta.kubernetes.io/instance-type":{},"f:beta.kubernetes.io/os":{},"f:cloud.google.com/metadata-proxy-ready":{},"f:failure-domain.beta.kubernetes.io/region":{},"f:failure-domain.beta.kubernetes.io/zone":{},"f:kubernetes.io/arch":{},"f:kubernetes.io/hostname":{},"f:kubernetes.io/os":{},"f:node.kubernetes.io/instance-type":{},"f:topology.kubernetes.io/region":{},"f:topology.kubernetes.io/zone":{}}},"f:spec":{"f:providerID":{}}} } {kube-controller-manager Update v1 2022-11-26 05:05:23 +0000 UTC FieldsV1 {"f:metadata":{"f:annotations":{"f:node.alpha.kubernetes.io/ttl":{}}},"f:spec":{"f:podCIDR":{},"f:podCIDRs":{".":{},"v:\"10.64.2.0/24\"":{}}}} } {kube-controller-manager Update v1 2022-11-26 05:13:31 +0000 UTC FieldsV1 {"f:status":{"f:conditions":{"k:{\"type\":\"NetworkUnavailable\"}":{"f:lastHeartbeatTime":{},"f:lastTransitionTime":{},"f:message":{},"f:reason":{},"f:status":{}}},"f:volumesAttached":{}}} status} {kubelet Update v1 2022-11-26 05:20:27 +0000 UTC FieldsV1 {"f:metadata":{"f:annotations":{"f:csi.volume.kubernetes.io/nodeid":{}},"f:labels":{"f:topology.hostpath.csi/node":{}}},"f:status":{"f:conditions":{"k:{\"type\":\"DiskPressure\"}":{"f:lastHeartbeatTime":{}},"k:{\"type\":\"MemoryPressure\"}":{"f:lastHeartbeatTime":{}},"k:{\"type\":\"PIDPressure\"}":{"f:lastHeartbeatTime":{}},"k:{\"type\":\"Ready\"}":{"f:lastHeartbeatTime":{},"f:lastTransitionTime":{},"f:message":{},"f:reason":{},"f:status":{}}},"f:images":{},"f:volumesInUse":{}}} status} {node-problem-detector Update v1 2022-11-26 05:20:30 +0000 UTC FieldsV1 {"f:status":{"f:conditions":{"k:{\"type\":\"CorruptDockerOverlay2\"}":{".":{},"f:lastHeartbeatTime":{},"f:lastTransitionTime":{},"f:message":{},"f:reason":{},"f:status":{},"f:type":{}},"k:{\"type\":\"FrequentContainerdRestart\"}":{".":{},"f:lastHeartbeatTime":{},"f:lastTransitionTime":{},"f:message":{},"f:reason":{},"f:status":{},"f:type":{}},"k:{\"type\":\"FrequentDockerRestart\"}":{".":{},"f:lastHeartbeatTime":{},"f:lastTransitionTime":{},"f:message":{},"f:reason":{},"f:status":{},"f:type":{}},"k:{\"type\":\"FrequentKubeletRestart\"}":{".":{},"f:lastHeartbeatTime":{},"f:lastTransitionTime":{},"f:message":{},"f:reason":{},"f:status":{},"f:type":{}},"k:{\"type\":\"FrequentUnregisterNetDevice\"}":{".":{},"f:lastHeartbeatTime":{},"f:lastTransitionTime":{},"f:message":{},"f:reason":{},"f:status":{},"f:type":{}},"k:{\"type\":\"KernelDeadlock\"}":{".":{},"f:lastHeartbeatTime":{},"f:lastTransitionTime":{},"f:message":{},"f:reason":{},"f:status":{},"f:type":{}},"k:{\"type\":\"ReadonlyFilesystem\"}":{".":{},"f:lastHeartbeatTime":{},"f:lastTransitionTime":{},"f:message":{},"f:reason":{},"f:status":{},"f:type":{}}}}} status}]},Spec:NodeSpec{PodCIDR:10.64.2.0/24,DoNotUseExternalID:,ProviderID:gce://k8s-jkns-e2e-gce-ci-reboot/us-west1-b/bootstrap-e2e-minion-group-0975,Unschedulable:false,Taints:[]Taint{},ConfigSource:nil,PodCIDRs:[10.64.2.0/24],},Status:NodeStatus{Capacity:ResourceList{attachable-volumes-gce-pd: {{127 0} {<nil>} 127 DecimalSI},cpu: {{2 0} {<nil>} 2 DecimalSI},ephemeral-storage: {{101203873792 0} {<nil>} 98831908Ki BinarySI},hugepages-1Gi: {{0 0} {<nil>} 0 DecimalSI},hugepages-2Mi: {{0 0} {<nil>} 0 DecimalSI},memory: {{7815430144 0} {<nil>} BinarySI},pods: {{110 0} {<nil>} 110 DecimalSI},},Allocatable:ResourceList{attachable-volumes-gce-pd: {{127 0} {<nil>} 127 DecimalSI},cpu: {{2 0} {<nil>} 2 DecimalSI},ephemeral-storage: {{91083486262 0} {<nil>} 91083486262 DecimalSI},hugepages-1Gi: {{0 0} {<nil>} 0 DecimalSI},hugepages-2Mi: {{0 0} {<nil>} 0 DecimalSI},memory: {{7553286144 0} {<nil>} BinarySI},pods: {{110 0} {<nil>} 110 DecimalSI},},Phase:,Conditions:[]NodeCondition{NodeCondition{Type:FrequentContainerdRestart,Status:False,LastHeartbeatTime:2022-11-26 05:20:30 +0000 UTC,LastTransitionTime:2022-11-26 05:05:27 +0000 UTC,Reason:NoFrequentContainerdRestart,Message:containerd is functioning properly,},NodeCondition{Type:KernelDeadlock,Status:False,LastHeartbeatTime:2022-11-26 05:20:30 +0000 UTC,LastTransitionTime:2022-11-26 05:05:27 +0000 UTC,Reason:KernelHasNoDeadlock,Message:kernel has no deadlock,},NodeCondition{Type:ReadonlyFilesystem,Status:False,LastHeartbeatTime:2022-11-26 05:20:30 +0000 UTC,LastTransitionTime:2022-11-26 05:05:27 +0000 UTC,Reason:FilesystemIsNotReadOnly,Message:Filesystem is not read-only,},NodeCondition{Type:CorruptDockerOverlay2,Status:False,LastHeartbeatTime:2022-11-26 05:20:30 +0000 UTC,LastTransitionTime:2022-11-26 05:05:27 +0000 UTC,Reason:NoCorruptDockerOverlay2,Message:docker overlay2 is functioning properly,},NodeCondition{Type:FrequentUnregisterNetDevice,Status:False,LastHeartbeatTime:2022-11-26 05:20:30 +0000 UTC,LastTransitionTime:2022-11-26 05:05:27 +0000 UTC,Reason:NoFrequentUnregisterNetDevice,Message:node is functioning properly,},NodeCondition{Type:FrequentKubeletRestart,Status:False,LastHeartbeatTime:2022-11-26 05:20:30 +0000 UTC,LastTransitionTime:2022-11-26 05:05:27 +0000 UTC,Reason:NoFrequentKubeletRestart,Message:kubelet is functioning properly,},NodeCondition{Type:FrequentDockerRestart,Status:False,LastHeartbeatTime:2022-11-26 05:20:30 +0000 UTC,LastTransitionTime:2022-11-26 05:05:27 +0000 UTC,Reason:NoFrequentDockerRestart,Message:docker is functioning properly,},NodeCondition{Type:NetworkUnavailable,Status:False,LastHeartbeatTime:2022-11-26 05:05:39 +0000 UTC,LastTransitionTime:2022-11-26 05:05:39 +0000 UTC,Reason:RouteCreated,Message:RouteController created a route,},NodeCondition{Type:MemoryPressure,Status:False,LastHeartbeatTime:2022-11-26 05:18:25 +0000 UTC,LastTransitionTime:2022-11-26 05:05:22 +0000 UTC,Reason:KubeletHasSufficientMemory,Message:kubelet has sufficient memory available,},NodeCondition{Type:DiskPressure,Status:False,LastHeartbeatTime:2022-11-26 05:18:25 +0000 UTC,LastTransitionTime:2022-11-26 05:05:22 +0000 UTC,Reason:KubeletHasNoDiskPressure,Message:kubelet has no disk pressure,},NodeCondition{Type:PIDPressure,Status:False,LastHeartbeatTime:2022-11-26 05:18:25 +0000 UTC,LastTransitionTime:2022-11-26 05:05:22 +0000 UTC,Reason:KubeletHasSufficientPID,Message:kubelet has sufficient PID available,},NodeCondition{Type:Ready,Status:True,LastHeartbeatTime:2022-11-26 05:18:25 +0000 UTC,LastTransitionTime:2022-11-26 05:05:23 +0000 UTC,Reason:KubeletReady,Message:kubelet is posting ready status. AppArmor enabled,},},Addresses:[]NodeAddress{NodeAddress{Type:InternalIP,Address:10.138.0.5,},NodeAddress{Type:ExternalIP,Address:35.247.104.236,},NodeAddress{Type:InternalDNS,Address:bootstrap-e2e-minion-group-0975.c.k8s-jkns-e2e-gce-ci-reboot.internal,},NodeAddress{Type:Hostname,Address:bootstrap-e2e-minion-group-0975.c.k8s-jkns-e2e-gce-ci-reboot.internal,},},DaemonEndpoints:NodeDaemonEndpoints{KubeletEndpoint:DaemonEndpoint{Port:10250,},},NodeInfo:NodeSystemInfo{MachineID:b38385f7976a474936da82f5f565760a,SystemUUID:b38385f7-976a-4749-36da-82f5f565760a,BootID:0aa74273-9b8d-48ac-bad0-0a1805f010ea,KernelVersion:5.10.123+,OSImage:Container-Optimized OS from Google,ContainerRuntimeVersion:containerd://1.7.0-beta.0-149-gd06318622,KubeletVersion:v1.27.0-alpha.0.50+70617042976dc1,KubeProxyVersion:v1.27.0-alpha.0.50+70617042976dc1,OperatingSystem:linux,Architecture:amd64,},Images:[]ContainerImage{ContainerImage{Names:[registry.k8s.io/e2e-test-images/jessie-dnsutils@sha256:24aaf2626d6b27864c29de2097e8bbb840b3a414271bf7c8995e431e47d8408e registry.k8s.io/e2e-test-images/jessie-dnsutils:1.7],SizeBytes:112030336,},ContainerImage{Names:[registry.k8s.io/sig-storage/nfs-provisioner@sha256:e943bb77c7df05ebdc8c7888b2db289b13bf9f012d6a3a5a74f14d4d5743d439 registry.k8s.io/sig-storage/nfs-provisioner:v3.0.1],SizeBytes:90632047,},ContainerImage{Names:[registry.k8s.io/kube-proxy-amd64:v1.27.0-alpha.0.50_70617042976dc1],SizeBytes:67201736,},ContainerImage{Names:[registry.k8s.io/e2e-test-images/agnhost@sha256:16bbf38c463a4223d8cfe4da12bc61010b082a79b4bb003e2d3ba3ece5dd5f9e registry.k8s.io/e2e-test-images/agnhost:2.43],SizeBytes:51706353,},ContainerImage{Names:[gke.gcr.io/prometheus-to-sd@sha256:e739643c3939ba0b161425f45a1989eedfc4a3b166db9a7100863296b4c70510 gke.gcr.io/prometheus-to-sd:v0.11.1-gke.1],SizeBytes:48742566,},ContainerImage{Names:[registry.k8s.io/e2e-test-images/httpd@sha256:148b022f5c5da426fc2f3c14b5c0867e58ef05961510c84749ac1fddcb0fef22 registry.k8s.io/e2e-test-images/httpd:2.4.38-4],SizeBytes:40764257,},ContainerImage{Names:[registry.k8s.io/metrics-server/metrics-server@sha256:6385aec64bb97040a5e692947107b81e178555c7a5b71caa90d733e4130efc10 registry.k8s.io/metrics-server/metrics-server:v0.5.2],SizeBytes:26023008,},ContainerImage{Names:[registry.k8s.io/sig-storage/csi-provisioner@sha256:ee3b525d5b89db99da3b8eb521d9cd90cb6e9ef0fbb651e98bb37be78d36b5b8 registry.k8s.io/sig-storage/csi-provisioner:v3.3.0],SizeBytes:25491225,},ContainerImage{Names:[registry.k8s.io/sig-storage/csi-resizer@sha256:425d8f1b769398127767b06ed97ce62578a3179bcb99809ce93a1649e025ffe7 registry.k8s.io/sig-storage/csi-resizer:v1.6.0],SizeBytes:24148884,},ContainerImage{Names:[registry.k8s.io/sig-storage/csi-snapshotter@sha256:291334908ddf71a4661fd7f6d9d97274de8a5378a2b6fdfeb2ce73414a34f82f registry.k8s.io/sig-storage/csi-snapshotter:v6.1.0],SizeBytes:23881995,},ContainerImage{Names:[registry.k8s.io/sig-storage/csi-attacher@sha256:9a685020911e2725ad019dbce6e4a5ab93d51e3d4557f115e64343345e05781b registry.k8s.io/sig-storage/csi-attacher:v4.0.0],SizeBytes:23847201,},ContainerImage{Names:[registry.k8s.io/sig-storage/hostpathplugin@sha256:92257881c1d6493cf18299a24af42330f891166560047902b8d431fb66b01af5 registry.k8s.io/sig-storage/hostpathplugin:v1.9.0],SizeBytes:18758628,},ContainerImage{Names:[registry.k8s.io/autoscaling/addon-resizer@sha256:43f129b81d28f0fdd54de6d8e7eacd5728030782e03db16087fc241ad747d3d6 registry.k8s.io/autoscaling/addon-resizer:1.8.14],SizeBytes:10153852,},ContainerImage{Names:[registry.k8s.io/sig-storage/csi-node-driver-registrar@sha256:0103eee7c35e3e0b5cd8cdca9850dc71c793cdeb6669d8be7a89440da2d06ae4 registry.k8s.io/sig-storage/csi-node-driver-registrar:v2.5.1],SizeBytes:9133109,},ContainerImage{Names:[registry.k8s.io/sig-storage/livenessprobe@sha256:933940f13b3ea0abc62e656c1aa5c5b47c04b15d71250413a6b821bd0c58b94e registry.k8s.io/sig-storage/livenessprobe:v2.7.0],SizeBytes:8688564,},ContainerImage{Names:[registry.k8s.io/kas-network-proxy/proxy-agent@sha256:48f2a4ec3e10553a81b8dd1c6fa5fe4bcc9617f78e71c1ca89c6921335e2d7da registry.k8s.io/kas-network-proxy/proxy-agent:v0.0.33],SizeBytes:8512162,},ContainerImage{Names:[registry.k8s.io/metadata-proxy@sha256:e914645f22e946bce5165737e1b244e0a296ad1f0f81a9531adc57af2780978a registry.k8s.io/metadata-proxy:v0.1.12],SizeBytes:5301657,},ContainerImage{Names:[registry.k8s.io/e2e-test-images/busybox@sha256:c318242786b139d18676b1c09a0ad7f15fc17f8f16a5b2e625cd0dc8c9703daf registry.k8s.io/e2e-test-images/busybox:1.29-2],SizeBytes:732424,},ContainerImage{Names:[registry.k8s.io/e2e-test-images/busybox@sha256:2e0f836850e09b8b7cc937681d6194537a09fbd5f6b9e08f4d646a85128e8937 registry.k8s.io/e2e-test-images/busybox:1.29-4],SizeBytes:731990,},ContainerImage{Names:[registry.k8s.io/pause@sha256:9001185023633d17a2f98ff69b6ff2615b8ea02a825adffa40422f51dfdcde9d registry.k8s.io/pause:3.8],SizeBytes:311286,},},VolumesInUse:[kubernetes.io/csi/csi-hostpath-multivolume-702^02fad0b0-6d49-11ed-9d4c-26ec58c28662],VolumesAttached:[]AttachedVolume{AttachedVolume{Name:kubernetes.io/csi/csi-hostpath-multivolume-702^02fad0b0-6d49-11ed-9d4c-26ec58c28662,DevicePath:,},},Config:nil,},} Nov 26 05:20:57.465: INFO: Logging kubelet events for node bootstrap-e2e-minion-group-0975 Nov 26 05:20:57.513: INFO: Logging pods the kubelet thinks is on node bootstrap-e2e-minion-group-0975 Nov 26 05:20:57.733: INFO: metrics-server-v0.5.2-867b8754b9-27xcg started at 2022-11-26 05:05:51 +0000 UTC (0+2 container statuses recorded) Nov 26 05:20:57.733: INFO: Container metrics-server ready: false, restart count 6 Nov 26 05:20:57.733: INFO: Container metrics-server-nanny ready: false, restart count 7 Nov 26 05:20:57.733: INFO: test-hostpath-type-2p74k started at 2022-11-26 05:17:26 +0000 UTC (0+1 container statuses recorded) Nov 26 05:20:57.733: INFO: Container host-path-testing ready: false, restart count 0 Nov 26 05:20:57.733: INFO: net-tiers-svc-sj7rf started at 2022-11-26 05:17:11 +0000 UTC (0+1 container statuses recorded) Nov 26 05:20:57.733: INFO: Container netexec ready: true, restart count 0 Nov 26 05:20:57.733: INFO: hostexec-bootstrap-e2e-minion-group-0975-vp665 started at 2022-11-26 05:17:32 +0000 UTC (0+1 container statuses recorded) Nov 26 05:20:57.733: INFO: Container agnhost-container ready: true, restart count 1 Nov 26 05:20:57.733: INFO: lb-internal-r5l4p started at 2022-11-26 05:16:06 +0000 UTC (0+1 container statuses recorded) Nov 26 05:20:57.733: INFO: Container netexec ready: true, restart count 0 Nov 26 05:20:57.733: INFO: test-hostpath-type-pk9r4 started at 2022-11-26 05:17:25 +0000 UTC (0+1 container statuses recorded) Nov 26 05:20:57.733: INFO: Container host-path-sh-testing ready: false, restart count 0 Nov 26 05:20:57.733: INFO: csi-mockplugin-0 started at 2022-11-26 05:17:37 +0000 UTC (0+4 container statuses recorded) Nov 26 05:20:57.733: INFO: Container busybox ready: true, restart count 2 Nov 26 05:20:57.733: INFO: Container csi-provisioner ready: true, restart count 2 Nov 26 05:20:57.733: INFO: Container driver-registrar ready: true, restart count 2 Nov 26 05:20:57.733: INFO: Container mock ready: true, restart count 2 Nov 26 05:20:57.733: INFO: netserver-0 started at 2022-11-26 05:09:51 +0000 UTC (0+1 container statuses recorded) Nov 26 05:20:57.733: INFO: Container webserver ready: false, restart count 5 Nov 26 05:20:57.733: INFO: csi-hostpathplugin-0 started at 2022-11-26 05:09:53 +0000 UTC (0+7 container statuses recorded) Nov 26 05:20:57.733: INFO: Container csi-attacher ready: true, restart count 3 Nov 26 05:20:57.733: INFO: Container csi-provisioner ready: true, restart count 3 Nov 26 05:20:57.733: INFO: Container csi-resizer ready: true, restart count 3 Nov 26 05:20:57.733: INFO: Container csi-snapshotter ready: true, restart count 3 Nov 26 05:20:57.733: INFO: Container hostpath ready: true, restart count 3 Nov 26 05:20:57.733: INFO: Container liveness-probe ready: true, restart count 3 Nov 26 05:20:57.733: INFO: Container node-driver-registrar ready: true, restart count 3 Nov 26 05:20:57.733: INFO: external-local-update-5xshb started at 2022-11-26 05:11:23 +0000 UTC (0+1 container statuses recorded) Nov 26 05:20:57.733: INFO: Container netexec ready: true, restart count 6 Nov 26 05:20:57.733: INFO: host-test-container-pod started at 2022-11-26 05:12:10 +0000 UTC (0+1 container statuses recorded) Nov 26 05:20:57.733: INFO: Container agnhost-container ready: true, restart count 3 Nov 26 05:20:57.733: INFO: pod-db90ef24-4c1c-4a0d-81ed-e023c5963392 started at 2022-11-26 05:13:08 +0000 UTC (0+1 container statuses recorded) Nov 26 05:20:57.733: INFO: Container write-pod ready: false, restart count 0 Nov 26 05:20:57.733: INFO: netserver-0 started at 2022-11-26 05:16:25 +0000 UTC (0+1 container statuses recorded) Nov 26 05:20:57.733: INFO: Container webserver ready: false, restart count 1 Nov 26 05:20:57.733: INFO: metadata-proxy-v0.1-mznrc started at 2022-11-26 05:05:24 +0000 UTC (0+2 container statuses recorded) Nov 26 05:20:57.733: INFO: Container metadata-proxy ready: true, restart count 0 Nov 26 05:20:57.733: INFO: Container prometheus-to-sd-exporter ready: true, restart count 0 Nov 26 05:20:57.733: INFO: konnectivity-agent-wfc29 started at 2022-11-26 05:05:39 +0000 UTC (0+1 container statuses recorded) Nov 26 05:20:57.733: INFO: Container konnectivity-agent ready: false, restart count 6 Nov 26 05:20:57.733: INFO: netserver-0 started at 2022-11-26 05:11:06 +0000 UTC (0+1 container statuses recorded) Nov 26 05:20:57.733: INFO: Container webserver ready: false, restart count 4 Nov 26 05:20:57.733: INFO: hostexec-bootstrap-e2e-minion-group-0975-rg9hf started at 2022-11-26 05:17:08 +0000 UTC (0+1 container statuses recorded) Nov 26 05:20:57.733: INFO: Container agnhost-container ready: true, restart count 2 Nov 26 05:20:57.733: INFO: csi-hostpathplugin-0 started at 2022-11-26 05:12:10 +0000 UTC (0+7 container statuses recorded) Nov 26 05:20:57.733: INFO: Container csi-attacher ready: true, restart count 5 Nov 26 05:20:57.733: INFO: Container csi-provisioner ready: true, restart count 5 Nov 26 05:20:57.733: INFO: Container csi-resizer ready: true, restart count 5 Nov 26 05:20:57.733: INFO: Container csi-snapshotter ready: true, restart count 5 Nov 26 05:20:57.733: INFO: Container hostpath ready: true, restart count 5 Nov 26 05:20:57.733: INFO: Container liveness-probe ready: true, restart count 5 Nov 26 05:20:57.733: INFO: Container node-driver-registrar ready: true, restart count 5 Nov 26 05:20:57.733: INFO: pod-secrets-ae6e2110-abaa-4433-ab33-c417114f813f started at 2022-11-26 05:10:35 +0000 UTC (0+1 container statuses recorded) Nov 26 05:20:57.733: INFO: Container creates-volume-test ready: false, restart count 0 Nov 26 05:20:57.733: INFO: csi-hostpathplugin-0 started at 2022-11-26 05:13:04 +0000 UTC (0+7 container statuses recorded) Nov 26 05:20:57.733: INFO: Container csi-attacher ready: true, restart count 0 Nov 26 05:20:57.733: INFO: Container csi-provisioner ready: true, restart count 0 Nov 26 05:20:57.733: INFO: Container csi-resizer ready: true, restart count 0 Nov 26 05:20:57.733: INFO: Container csi-snapshotter ready: true, restart count 0 Nov 26 05:20:57.733: INFO: Container hostpath ready: true, restart count 0 Nov 26 05:20:57.733: INFO: Container liveness-probe ready: true, restart count 0 Nov 26 05:20:57.733: INFO: Container node-driver-registrar ready: true, restart count 0 Nov 26 05:20:57.733: INFO: kube-proxy-bootstrap-e2e-minion-group-0975 started at 2022-11-26 05:05:23 +0000 UTC (0+1 container statuses recorded) Nov 26 05:20:57.733: INFO: Container kube-proxy ready: false, restart count 6 Nov 26 05:20:57.733: INFO: csi-hostpathplugin-0 started at 2022-11-26 05:13:03 +0000 UTC (0+7 container statuses recorded) Nov 26 05:20:57.733: INFO: Container csi-attacher ready: true, restart count 2 Nov 26 05:20:57.733: INFO: Container csi-provisioner ready: true, restart count 2 Nov 26 05:20:57.733: INFO: Container csi-resizer ready: true, restart count 2 Nov 26 05:20:57.733: INFO: Container csi-snapshotter ready: true, restart count 2 Nov 26 05:20:57.733: INFO: Container hostpath ready: true, restart count 2 Nov 26 05:20:57.733: INFO: Container liveness-probe ready: true, restart count 2 Nov 26 05:20:57.733: INFO: Container node-driver-registrar ready: true, restart count 2 Nov 26 05:20:57.733: INFO: netserver-0 started at 2022-11-26 05:11:29 +0000 UTC (0+1 container statuses recorded) Nov 26 05:20:57.733: INFO: Container webserver ready: true, restart count 5 Nov 26 05:20:58.512: INFO: Latency metrics for node bootstrap-e2e-minion-group-0975 Nov 26 05:20:58.512: INFO: Logging node info for node bootstrap-e2e-minion-group-rwsn Nov 26 05:20:58.571: INFO: Node Info: &Node{ObjectMeta:{bootstrap-e2e-minion-group-rwsn adedf225-b462-4305-a60c-9042d31ca1d2 8762 0 2022-11-26 05:05:15 +0000 UTC <nil> <nil> map[beta.kubernetes.io/arch:amd64 beta.kubernetes.io/instance-type:n1-standard-2 beta.kubernetes.io/os:linux cloud.google.com/metadata-proxy-ready:true failure-domain.beta.kubernetes.io/region:us-west1 failure-domain.beta.kubernetes.io/zone:us-west1-b kubernetes.io/arch:amd64 kubernetes.io/hostname:bootstrap-e2e-minion-group-rwsn kubernetes.io/os:linux node.kubernetes.io/instance-type:n1-standard-2 topology.hostpath.csi/node:bootstrap-e2e-minion-group-rwsn topology.kubernetes.io/region:us-west1 topology.kubernetes.io/zone:us-west1-b] map[csi.volume.kubernetes.io/nodeid:{"csi-hostpath-multivolume-4683":"bootstrap-e2e-minion-group-rwsn","csi-hostpath-multivolume-6872":"bootstrap-e2e-minion-group-rwsn"} node.alpha.kubernetes.io/ttl:0 volumes.kubernetes.io/controller-managed-attach-detach:true] [] [] [{kubelet Update v1 2022-11-26 05:05:15 +0000 UTC FieldsV1 {"f:metadata":{"f:annotations":{".":{},"f:volumes.kubernetes.io/controller-managed-attach-detach":{}},"f:labels":{".":{},"f:beta.kubernetes.io/arch":{},"f:beta.kubernetes.io/instance-type":{},"f:beta.kubernetes.io/os":{},"f:cloud.google.com/metadata-proxy-ready":{},"f:failure-domain.beta.kubernetes.io/region":{},"f:failure-domain.beta.kubernetes.io/zone":{},"f:kubernetes.io/arch":{},"f:kubernetes.io/hostname":{},"f:kubernetes.io/os":{},"f:node.kubernetes.io/instance-type":{},"f:topology.kubernetes.io/region":{},"f:topology.kubernetes.io/zone":{}}},"f:spec":{"f:providerID":{}}} } {kube-controller-manager Update v1 2022-11-26 05:05:17 +0000 UTC FieldsV1 {"f:metadata":{"f:annotations":{"f:node.alpha.kubernetes.io/ttl":{}}},"f:spec":{"f:podCIDR":{},"f:podCIDRs":{".":{},"v:\"10.64.0.0/24\"":{}}}} } {kube-controller-manager Update v1 2022-11-26 05:16:30 +0000 UTC FieldsV1 {"f:status":{"f:conditions":{"k:{\"type\":\"NetworkUnavailable\"}":{"f:lastHeartbeatTime":{},"f:lastTransitionTime":{},"f:message":{},"f:reason":{},"f:status":{}}},"f:volumesAttached":{}}} status} {node-problem-detector Update v1 2022-11-26 05:20:21 +0000 UTC FieldsV1 {"f:status":{"f:conditions":{"k:{\"type\":\"CorruptDockerOverlay2\"}":{".":{},"f:lastHeartbeatTime":{},"f:lastTransitionTime":{},"f:message":{},"f:reason":{},"f:status":{},"f:type":{}},"k:{\"type\":\"FrequentContainerdRestart\"}":{".":{},"f:lastHeartbeatTime":{},"f:lastTransitionTime":{},"f:message":{},"f:reason":{},"f:status":{},"f:type":{}},"k:{\"type\":\"FrequentDockerRestart\"}":{".":{},"f:lastHeartbeatTime":{},"f:lastTransitionTime":{},"f:message":{},"f:reason":{},"f:status":{},"f:type":{}},"k:{\"type\":\"FrequentKubeletRestart\"}":{".":{},"f:lastHeartbeatTime":{},"f:lastTransitionTime":{},"f:message":{},"f:reason":{},"f:status":{},"f:type":{}},"k:{\"type\":\"FrequentUnregisterNetDevice\"}":{".":{},"f:lastHeartbeatTime":{},"f:lastTransitionTime":{},"f:message":{},"f:reason":{},"f:status":{},"f:type":{}},"k:{\"type\":\"KernelDeadlock\"}":{".":{},"f:lastHeartbeatTime":{},"f:lastTransitionTime":{},"f:message":{},"f:reason":{},"f:status":{},"f:type":{}},"k:{\"type\":\"ReadonlyFilesystem\"}":{".":{},"f:lastHeartbeatTime":{},"f:lastTransitionTime":{},"f:message":{},"f:reason":{},"f:status":{},"f:type":{}}}}} status} {kubelet Update v1 2022-11-26 05:20:35 +0000 UTC FieldsV1 {"f:metadata":{"f:annotations":{"f:csi.volume.kubernetes.io/nodeid":{}},"f:labels":{"f:topology.hostpath.csi/node":{}}},"f:status":{"f:conditions":{"k:{\"type\":\"DiskPressure\"}":{"f:lastHeartbeatTime":{}},"k:{\"type\":\"MemoryPressure\"}":{"f:lastHeartbeatTime":{}},"k:{\"type\":\"PIDPressure\"}":{"f:lastHeartbeatTime":{}},"k:{\"type\":\"Ready\"}":{"f:lastHeartbeatTime":{},"f:lastTransitionTime":{},"f:message":{},"f:reason":{},"f:status":{}}},"f:images":{},"f:volumesInUse":{}}} status}]},Spec:NodeSpec{PodCIDR:10.64.0.0/24,DoNotUseExternalID:,ProviderID:gce://k8s-jkns-e2e-gce-ci-reboot/us-west1-b/bootstrap-e2e-minion-group-rwsn,Unschedulable:false,Taints:[]Taint{},ConfigSource:nil,PodCIDRs:[10.64.0.0/24],},Status:NodeStatus{Capacity:ResourceList{attachable-volumes-gce-pd: {{127 0} {<nil>} 127 DecimalSI},cpu: {{2 0} {<nil>} 2 DecimalSI},ephemeral-storage: {{101203873792 0} {<nil>} 98831908Ki BinarySI},hugepages-1Gi: {{0 0} {<nil>} 0 DecimalSI},hugepages-2Mi: {{0 0} {<nil>} 0 DecimalSI},memory: {{7815438336 0} {<nil>} BinarySI},pods: {{110 0} {<nil>} 110 DecimalSI},},Allocatable:ResourceList{attachable-volumes-gce-pd: {{127 0} {<nil>} 127 DecimalSI},cpu: {{2 0} {<nil>} 2 DecimalSI},ephemeral-storage: {{91083486262 0} {<nil>} 91083486262 DecimalSI},hugepages-1Gi: {{0 0} {<nil>} 0 DecimalSI},hugepages-2Mi: {{0 0} {<nil>} 0 DecimalSI},memory: {{7553294336 0} {<nil>} BinarySI},pods: {{110 0} {<nil>} 110 DecimalSI},},Phase:,Conditions:[]NodeCondition{NodeCondition{Type:FrequentContainerdRestart,Status:False,LastHeartbeatTime:2022-11-26 05:20:21 +0000 UTC,LastTransitionTime:2022-11-26 05:05:18 +0000 UTC,Reason:NoFrequentContainerdRestart,Message:containerd is functioning properly,},NodeCondition{Type:KernelDeadlock,Status:False,LastHeartbeatTime:2022-11-26 05:20:21 +0000 UTC,LastTransitionTime:2022-11-26 05:05:18 +0000 UTC,Reason:KernelHasNoDeadlock,Message:kernel has no deadlock,},NodeCondition{Type:ReadonlyFilesystem,Status:False,LastHeartbeatTime:2022-11-26 05:20:21 +0000 UTC,LastTransitionTime:2022-11-26 05:05:18 +0000 UTC,Reason:FilesystemIsNotReadOnly,Message:Filesystem is not read-only,},NodeCondition{Type:CorruptDockerOverlay2,Status:False,LastHeartbeatTime:2022-11-26 05:20:21 +0000 UTC,LastTransitionTime:2022-11-26 05:05:18 +0000 UTC,Reason:NoCorruptDockerOverlay2,Message:docker overlay2 is functioning properly,},NodeCondition{Type:FrequentUnregisterNetDevice,Status:False,LastHeartbeatTime:2022-11-26 05:20:21 +0000 UTC,LastTransitionTime:2022-11-26 05:05:18 +0000 UTC,Reason:NoFrequentUnregisterNetDevice,Message:node is functioning properly,},NodeCondition{Type:FrequentKubeletRestart,Status:False,LastHeartbeatTime:2022-11-26 05:20:21 +0000 UTC,LastTransitionTime:2022-11-26 05:05:18 +0000 UTC,Reason:NoFrequentKubeletRestart,Message:kubelet is functioning properly,},NodeCondition{Type:FrequentDockerRestart,Status:False,LastHeartbeatTime:2022-11-26 05:20:21 +0000 UTC,LastTransitionTime:2022-11-26 05:05:18 +0000 UTC,Reason:NoFrequentDockerRestart,Message:docker is functioning properly,},NodeCondition{Type:NetworkUnavailable,Status:False,LastHeartbeatTime:2022-11-26 05:05:25 +0000 UTC,LastTransitionTime:2022-11-26 05:05:25 +0000 UTC,Reason:RouteCreated,Message:RouteController created a route,},NodeCondition{Type:MemoryPressure,Status:False,LastHeartbeatTime:2022-11-26 05:17:07 +0000 UTC,LastTransitionTime:2022-11-26 05:05:15 +0000 UTC,Reason:KubeletHasSufficientMemory,Message:kubelet has sufficient memory available,},NodeCondition{Type:DiskPressure,Status:False,LastHeartbeatTime:2022-11-26 05:17:07 +0000 UTC,LastTransitionTime:2022-11-26 05:05:15 +0000 UTC,Reason:KubeletHasNoDiskPressure,Message:kubelet has no disk pressure,},NodeCondition{Type:PIDPressure,Status:False,LastHeartbeatTime:2022-11-26 05:17:07 +0000 UTC,LastTransitionTime:2022-11-26 05:05:15 +0000 UTC,Reason:KubeletHasSufficientPID,Message:kubelet has sufficient PID available,},NodeCondition{Type:Ready,Status:True,LastHeartbeatTime:2022-11-26 05:17:07 +0000 UTC,LastTransitionTime:2022-11-26 05:05:17 +0000 UTC,Reason:KubeletReady,Message:kubelet is posting ready status. AppArmor enabled,},},Addresses:[]NodeAddress{NodeAddress{Type:InternalIP,Address:10.138.0.4,},NodeAddress{Type:ExternalIP,Address:35.230.27.45,},NodeAddress{Type:InternalDNS,Address:bootstrap-e2e-minion-group-rwsn.c.k8s-jkns-e2e-gce-ci-reboot.internal,},NodeAddress{Type:Hostname,Address:bootstrap-e2e-minion-group-rwsn.c.k8s-jkns-e2e-gce-ci-reboot.internal,},},DaemonEndpoints:NodeDaemonEndpoints{KubeletEndpoint:DaemonEndpoint{Port:10250,},},NodeInfo:NodeSystemInfo{MachineID:69e22ce1281686d6cde1b3211b34b8af,SystemUUID:69e22ce1-2816-86d6-cde1-b3211b34b8af,BootID:79e9e1fd-2a39-42ec-b72b-72836d213862,KernelVersion:5.10.123+,OSImage:Container-Optimized OS from Google,ContainerRuntimeVersion:containerd://1.7.0-beta.0-149-gd06318622,KubeletVersion:v1.27.0-alpha.0.50+70617042976dc1,KubeProxyVersion:v1.27.0-alpha.0.50+70617042976dc1,OperatingSystem:linux,Architecture:amd64,},Images:[]ContainerImage{ContainerImage{Names:[registry.k8s.io/e2e-test-images/volume/nfs@sha256:3bda73f2428522b0e342af80a0b9679e8594c2126f2b3cca39ed787589741b9e registry.k8s.io/e2e-test-images/volume/nfs:1.3],SizeBytes:95836203,},ContainerImage{Names:[registry.k8s.io/sig-storage/nfs-provisioner@sha256:e943bb77c7df05ebdc8c7888b2db289b13bf9f012d6a3a5a74f14d4d5743d439 registry.k8s.io/sig-storage/nfs-provisioner:v3.0.1],SizeBytes:90632047,},ContainerImage{Names:[registry.k8s.io/kube-proxy-amd64:v1.27.0-alpha.0.50_70617042976dc1],SizeBytes:67201736,},ContainerImage{Names:[registry.k8s.io/e2e-test-images/agnhost@sha256:16bbf38c463a4223d8cfe4da12bc61010b082a79b4bb003e2d3ba3ece5dd5f9e registry.k8s.io/e2e-test-images/agnhost:2.43],SizeBytes:51706353,},ContainerImage{Names:[gke.gcr.io/prometheus-to-sd@sha256:e739643c3939ba0b161425f45a1989eedfc4a3b166db9a7100863296b4c70510 gke.gcr.io/prometheus-to-sd:v0.11.1-gke.1],SizeBytes:48742566,},ContainerImage{Names:[registry.k8s.io/metrics-server/metrics-server@sha256:6385aec64bb97040a5e692947107b81e178555c7a5b71caa90d733e4130efc10 registry.k8s.io/metrics-server/metrics-server:v0.5.2],SizeBytes:26023008,},ContainerImage{Names:[registry.k8s.io/sig-storage/csi-provisioner@sha256:ee3b525d5b89db99da3b8eb521d9cd90cb6e9ef0fbb651e98bb37be78d36b5b8 registry.k8s.io/sig-storage/csi-provisioner:v3.3.0],SizeBytes:25491225,},ContainerImage{Names:[registry.k8s.io/sig-storage/csi-resizer@sha256:425d8f1b769398127767b06ed97ce62578a3179bcb99809ce93a1649e025ffe7 registry.k8s.io/sig-storage/csi-resizer:v1.6.0],SizeBytes:24148884,},ContainerImage{Names:[registry.k8s.io/sig-storage/csi-snapshotter@sha256:291334908ddf71a4661fd7f6d9d97274de8a5378a2b6fdfeb2ce73414a34f82f registry.k8s.io/sig-storage/csi-snapshotter:v6.1.0],SizeBytes:23881995,},ContainerImage{Names:[registry.k8s.io/sig-storage/csi-attacher@sha256:9a685020911e2725ad019dbce6e4a5ab93d51e3d4557f115e64343345e05781b registry.k8s.io/sig-storage/csi-attacher:v4.0.0],SizeBytes:23847201,},ContainerImage{Names:[registry.k8s.io/sig-storage/snapshot-controller@sha256:823c75d0c45d1427f6d850070956d9ca657140a7bbf828381541d1d808475280 registry.k8s.io/sig-storage/snapshot-controller:v6.1.0],SizeBytes:22620891,},ContainerImage{Names:[registry.k8s.io/sig-storage/hostpathplugin@sha256:92257881c1d6493cf18299a24af42330f891166560047902b8d431fb66b01af5 registry.k8s.io/sig-storage/hostpathplugin:v1.9.0],SizeBytes:18758628,},ContainerImage{Names:[registry.k8s.io/cpa/cluster-proportional-autoscaler@sha256:fd636b33485c7826fb20ef0688a83ee0910317dbb6c0c6f3ad14661c1db25def registry.k8s.io/cpa/cluster-proportional-autoscaler:1.8.4],SizeBytes:15209393,},ContainerImage{Names:[registry.k8s.io/coredns/coredns@sha256:8e352a029d304ca7431c6507b56800636c321cb52289686a581ab70aaa8a2e2a registry.k8s.io/coredns/coredns:v1.9.3],SizeBytes:14837849,},ContainerImage{Names:[registry.k8s.io/autoscaling/addon-resizer@sha256:43f129b81d28f0fdd54de6d8e7eacd5728030782e03db16087fc241ad747d3d6 registry.k8s.io/autoscaling/addon-resizer:1.8.14],SizeBytes:10153852,},ContainerImage{Names:[registry.k8s.io/sig-storage/csi-node-driver-registrar@sha256:0103eee7c35e3e0b5cd8cdca9850dc71c793cdeb6669d8be7a89440da2d06ae4 registry.k8s.io/sig-storage/csi-node-driver-registrar:v2.5.1],SizeBytes:9133109,},ContainerImage{Names:[registry.k8s.io/sig-storage/livenessprobe@sha256:933940f13b3ea0abc62e656c1aa5c5b47c04b15d71250413a6b821bd0c58b94e registry.k8s.io/sig-storage/livenessprobe:v2.7.0],SizeBytes:8688564,},ContainerImage{Names:[registry.k8s.io/kas-network-proxy/proxy-agent@sha256:48f2a4ec3e10553a81b8dd1c6fa5fe4bcc9617f78e71c1ca89c6921335e2d7da registry.k8s.io/kas-network-proxy/proxy-agent:v0.0.33],SizeBytes:8512162,},ContainerImage{Names:[registry.k8s.io/networking/ingress-gce-404-server-with-metrics-amd64@sha256:7eb7b3cee4d33c10c49893ad3c386232b86d4067de5251294d4c620d6e072b93 registry.k8s.io/networking/ingress-gce-404-server-with-metrics-amd64:v1.10.11],SizeBytes:6463068,},ContainerImage{Names:[registry.k8s.io/metadata-proxy@sha256:e914645f22e946bce5165737e1b244e0a296ad1f0f81a9531adc57af2780978a registry.k8s.io/metadata-proxy:v0.1.12],SizeBytes:5301657,},ContainerImage{Names:[registry.k8s.io/e2e-test-images/busybox@sha256:c318242786b139d18676b1c09a0ad7f15fc17f8f16a5b2e625cd0dc8c9703daf registry.k8s.io/e2e-test-images/busybox:1.29-2],SizeBytes:732424,},ContainerImage{Names:[registry.k8s.io/e2e-test-images/busybox@sha256:2e0f836850e09b8b7cc937681d6194537a09fbd5f6b9e08f4d646a85128e8937 registry.k8s.io/e2e-test-images/busybox:1.29-4],SizeBytes:731990,},ContainerImage{Names:[registry.k8s.io/pause@sha256:7031c1b283388d2c2e09b57badb803c05ebed362dc88d84b480cc47f72a21097 registry.k8s.io/pause:3.9],SizeBytes:321520,},ContainerImage{Names:[registry.k8s.io/pause@sha256:9001185023633d17a2f98ff69b6ff2615b8ea02a825adffa40422f51dfdcde9d registry.k8s.io/pause:3.8],SizeBytes:311286,},},VolumesInUse:[kubernetes.io/csi/csi-hostpath-multivolume-4683^6fe7a3e3-6d49-11ed-ac34-8a6e303c97ed kubernetes.io/csi/csi-hostpath-multivolume-4683^7ad9a9cd-6d49-11ed-b2a0-f2f9c395d3b2 kubernetes.io/csi/csi-hostpath-multivolume-6872^8dfb6ea6-6d49-11ed-bbbd-7a3847645704 kubernetes.io/csi/csi-hostpath-provisioning-6984^3b0457df-6d48-11ed-9e4d-d626cdd9a50a],VolumesAttached:[]AttachedVolume{AttachedVolume{Name:kubernetes.io/csi/csi-hostpath-provisioning-6984^3b0457df-6d48-11ed-9e4d-d626cdd9a50a,DevicePath:,},AttachedVolume{Name:kubernetes.io/csi/csi-hostpath-multivolume-4683^7ad9a9cd-6d49-11ed-b2a0-f2f9c395d3b2,DevicePath:,},AttachedVolume{Name:kubernetes.io/csi/csi-hostpath-multivolume-4683^6fe7a3e3-6d49-11ed-ac34-8a6e303c97ed,DevicePath:,},},Config:nil,},} Nov 26 05:20:58.572: INFO: Logging kubelet events for node bootstrap-e2e-minion-group-rwsn Nov 26 05:20:58.616: INFO: Logging pods the kubelet thinks is on node bootstrap-e2e-minion-group-rwsn Nov 26 05:20:58.685: INFO: csi-mockplugin-0 started at 2022-11-26 05:13:01 +0000 UTC (0+4 container statuses recorded) Nov 26 05:20:58.685: INFO: Container busybox ready: false, restart count 4 Nov 26 05:20:58.685: INFO: Container csi-provisioner ready: false, restart count 4 Nov 26 05:20:58.685: INFO: Container driver-registrar ready: false, restart count 4 Nov 26 05:20:58.685: INFO: Container mock ready: false, restart count 4 Nov 26 05:20:58.685: INFO: hostexec-bootstrap-e2e-minion-group-rwsn-l47jh started at 2022-11-26 05:19:38 +0000 UTC (0+1 container statuses recorded) Nov 26 05:20:58.685: INFO: Container agnhost-container ready: true, restart count 0 Nov 26 05:20:58.685: INFO: csi-mockplugin-0 started at 2022-11-26 05:08:11 +0000 UTC (0+4 container statuses recorded) Nov 26 05:20:58.685: INFO: Container busybox ready: true, restart count 4 Nov 26 05:20:58.685: INFO: Container csi-provisioner ready: false, restart count 3 Nov 26 05:20:58.685: INFO: Container driver-registrar ready: false, restart count 4 Nov 26 05:20:58.685: INFO: Container mock ready: false, restart count 4 Nov 26 05:20:58.685: INFO: nfs-server started at 2022-11-26 05:13:35 +0000 UTC (0+1 container statuses recorded) Nov 26 05:20:58.685: INFO: Container nfs-server ready: true, restart count 2 Nov 26 05:20:58.685: INFO: netserver-1 started at 2022-11-26 05:11:29 +0000 UTC (0+1 container statuses recorded) Nov 26 05:20:58.685: INFO: Container webserver ready: true, restart count 5 Nov 26 05:20:58.685: INFO: l7-default-backend-8549d69d99-8cfzv started at 2022-11-26 05:05:25 +0000 UTC (0+1 container statuses recorded) Nov 26 05:20:58.685: INFO: Container default-http-backend ready: true, restart count 0 Nov 26 05:20:58.685: INFO: volume-snapshot-controller-0 started at 2022-11-26 05:05:25 +0000 UTC (0+1 container statuses recorded) Nov 26 05:20:58.685: INFO: Container volume-snapshot-controller ready: true, restart count 6 Nov 26 05:20:58.685: INFO: netserver-1 started at 2022-11-26 05:09:51 +0000 UTC (0+1 container statuses recorded) Nov 26 05:20:58.685: INFO: Container webserver ready: false, restart count 5 Nov 26 05:20:58.685: INFO: hostexec-bootstrap-e2e-minion-group-rwsn-z9rtd started at 2022-11-26 05:17:40 +0000 UTC (0+1 container statuses recorded) Nov 26 05:20:58.685: INFO: Container agnhost-container ready: true, restart count 1 Nov 26 05:20:58.685: INFO: csi-hostpathplugin-0 started at 2022-11-26 05:16:00 +0000 UTC (0+7 container statuses recorded) Nov 26 05:20:58.685: INFO: Container csi-attacher ready: true, restart count 5 Nov 26 05:20:58.685: INFO: Container csi-provisioner ready: true, restart count 5 Nov 26 05:20:58.685: INFO: Container csi-resizer ready: true, restart count 5 Nov 26 05:20:58.685: INFO: Container csi-snapshotter ready: true, restart count 5 Nov 26 05:20:58.685: INFO: Container hostpath ready: true, restart count 5 Nov 26 05:20:58.685: INFO: Container liveness-probe ready: true, restart count 5 Nov 26 05:20:58.685: INFO: Container node-driver-registrar ready: true, restart count 5 Nov 26 05:20:58.685: INFO: coredns-6d97d5ddb-xk6v4 started at 2022-11-26 05:05:25 +0000 UTC (0+1 container statuses recorded) Nov 26 05:20:58.685: INFO: Container coredns ready: false, restart count 7 Nov 26 05:20:58.685: INFO: pod-subpath-test-dynamicpv-h6m7 started at 2022-11-26 05:07:31 +0000 UTC (1+1 container statuses recorded) Nov 26 05:20:58.685: INFO: Init container init-volume-dynamicpv-h6m7 ready: true, restart count 0 Nov 26 05:20:58.685: INFO: Container test-container-subpath-dynamicpv-h6m7 ready: false, restart count 0 Nov 26 05:20:58.685: INFO: pod-04078049-bcc4-43bf-8abc-addcb96042db started at 2022-11-26 05:16:29 +0000 UTC (0+1 container statuses recorded) Nov 26 05:20:58.685: INFO: Container write-pod ready: false, restart count 0 Nov 26 05:20:58.685: INFO: kube-dns-autoscaler-5f6455f985-xktw8 started at 2022-11-26 05:05:25 +0000 UTC (0+1 container statuses recorded) Nov 26 05:20:58.685: INFO: Container autoscaler ready: false, restart count 6 Nov 26 05:20:58.685: INFO: coredns-6d97d5ddb-p2xpt started at 2022-11-26 05:05:31 +0000 UTC (0+1 container statuses recorded) Nov 26 05:20:58.685: INFO: Container coredns ready: true, restart count 7 Nov 26 05:20:58.685: INFO: netserver-1 started at 2022-11-26 05:16:25 +0000 UTC (0+1 container statuses recorded) Nov 26 05:20:58.685: INFO: Container webserver ready: true, restart count 5 Nov 26 05:20:58.685: INFO: pvc-volume-tester-5h2gx started at 2022-11-26 05:13:54 +0000 UTC (0+1 container statuses recorded) Nov 26 05:20:58.685: INFO: Container volume-tester ready: false, restart count 0 Nov 26 05:20:58.685: INFO: kube-proxy-bootstrap-e2e-minion-group-rwsn started at 2022-11-26 05:05:15 +0000 UTC (0+1 container statuses recorded) Nov 26 05:20:58.685: INFO: Container kube-proxy ready: false, restart count 5 Nov 26 05:20:58.685: INFO: csi-hostpathplugin-0 started at 2022-11-26 05:07:19 +0000 UTC (0+7 container statuses recorded) Nov 26 05:20:58.685: INFO: Container csi-attacher ready: false, restart count 7 Nov 26 05:20:58.685: INFO: Container csi-provisioner ready: false, restart count 7 Nov 26 05:20:58.685: INFO: Container csi-resizer ready: false, restart count 7 Nov 26 05:20:58.685: INFO: Container csi-snapshotter ready: false, restart count 7 Nov 26 05:20:58.685: INFO: Container hostpath ready: false, restart count 7 Nov 26 05:20:58.685: INFO: Container liveness-probe ready: false, restart count 7 Nov 26 05:20:58.685: INFO: Container node-driver-registrar ready: false, restart count 7 Nov 26 05:20:58.685: INFO: netserver-1 started at 2022-11-26 05:11:06 +0000 UTC (0+1 container statuses recorded) Nov 26 05:20:58.685: INFO: Container webserver ready: false, restart count 6 Nov 26 05:20:58.685: INFO: metadata-proxy-v0.1-qlckk started at 2022-11-26 05:05:16 +0000 UTC (0+2 container statuses recorded) Nov 26 05:20:58.685: INFO: Container metadata-proxy ready: true, restart count 0 Nov 26 05:20:58.685: INFO: Container prometheus-to-sd-exporter ready: true, restart count 0 Nov 26 05:20:58.685: INFO: csi-hostpathplugin-0 started at 2022-11-26 05:10:11 +0000 UTC (0+7 container statuses recorded) Nov 26 05:20:58.685: INFO: Container csi-attacher ready: false, restart count 6 Nov 26 05:20:58.685: INFO: Container csi-provisioner ready: false, restart count 6 Nov 26 05:20:58.685: INFO: Container csi-resizer ready: false, restart count 6 Nov 26 05:20:58.685: INFO: Container csi-snapshotter ready: false, restart count 6 Nov 26 05:20:58.685: INFO: Container hostpath ready: false, restart count 6 Nov 26 05:20:58.685: INFO: Container liveness-probe ready: false, restart count 6 Nov 26 05:20:58.685: INFO: Container node-driver-registrar ready: false, restart count 6 Nov 26 05:20:58.685: INFO: csi-hostpathplugin-0 started at 2022-11-26 05:16:04 +0000 UTC (0+7 container statuses recorded) Nov 26 05:20:58.685: INFO: Container csi-attacher ready: true, restart count 4 Nov 26 05:20:58.685: INFO: Container csi-provisioner ready: true, restart count 4 Nov 26 05:20:58.685: INFO: Container csi-resizer ready: true, restart count 4 Nov 26 05:20:58.685: INFO: Container csi-snapshotter ready: true, restart count 4 Nov 26 05:20:58.685: INFO: Container hostpath ready: true, restart count 4 Nov 26 05:20:58.685: INFO: Container liveness-probe ready: true, restart count 4 Nov 26 05:20:58.685: INFO: Container node-driver-registrar ready: true, restart count 4 Nov 26 05:20:58.685: INFO: konnectivity-agent-sxg5q started at 2022-11-26 05:05:25 +0000 UTC (0+1 container statuses recorded) Nov 26 05:20:58.685: INFO: Container konnectivity-agent ready: false, restart count 6 Nov 26 05:20:58.685: INFO: pod-ce08e819-7446-4933-9e24-3a40ac81db8b started at 2022-11-26 05:17:01 +0000 UTC (0+1 container statuses recorded) Nov 26 05:20:58.685: INFO: Container write-pod ready: false, restart count 0 Nov 26 05:20:59.037: INFO: Latency metrics for node bootstrap-e2e-minion-group-rwsn Nov 26 05:20:59.037: INFO: Logging node info for node bootstrap-e2e-minion-group-vhdj Nov 26 05:20:59.079: INFO: Node Info: &Node{ObjectMeta:{bootstrap-e2e-minion-group-vhdj 0d27fd16-9ec8-4428-bb1e-97fead6c7a66 9018 0 2022-11-26 05:05:23 +0000 UTC <nil> <nil> map[beta.kubernetes.io/arch:amd64 beta.kubernetes.io/instance-type:n1-standard-2 beta.kubernetes.io/os:linux cloud.google.com/metadata-proxy-ready:true failure-domain.beta.kubernetes.io/region:us-west1 failure-domain.beta.kubernetes.io/zone:us-west1-b kubernetes.io/arch:amd64 kubernetes.io/hostname:bootstrap-e2e-minion-group-vhdj kubernetes.io/os:linux node.kubernetes.io/instance-type:n1-standard-2 topology.hostpath.csi/node:bootstrap-e2e-minion-group-vhdj topology.kubernetes.io/region:us-west1 topology.kubernetes.io/zone:us-west1-b] map[csi.volume.kubernetes.io/nodeid:{"csi-hostpath-multivolume-4149":"bootstrap-e2e-minion-group-vhdj","csi-hostpath-provisioning-6216":"bootstrap-e2e-minion-group-vhdj","csi-hostpath-provisioning-6223":"bootstrap-e2e-minion-group-vhdj","csi-hostpath-provisioning-7617":"bootstrap-e2e-minion-group-vhdj","csi-hostpath-volumeio-4786":"bootstrap-e2e-minion-group-vhdj","csi-hostpath-volumemode-1700":"bootstrap-e2e-minion-group-vhdj"} node.alpha.kubernetes.io/ttl:0 volumes.kubernetes.io/controller-managed-attach-detach:true] [] [] [{kubelet Update v1 2022-11-26 05:05:23 +0000 UTC FieldsV1 {"f:metadata":{"f:annotations":{".":{},"f:volumes.kubernetes.io/controller-managed-attach-detach":{}},"f:labels":{".":{},"f:beta.kubernetes.io/arch":{},"f:beta.kubernetes.io/instance-type":{},"f:beta.kubernetes.io/os":{},"f:cloud.google.com/metadata-proxy-ready":{},"f:failure-domain.beta.kubernetes.io/region":{},"f:failure-domain.beta.kubernetes.io/zone":{},"f:kubernetes.io/arch":{},"f:kubernetes.io/hostname":{},"f:kubernetes.io/os":{},"f:node.kubernetes.io/instance-type":{},"f:topology.kubernetes.io/region":{},"f:topology.kubernetes.io/zone":{}}},"f:spec":{"f:providerID":{}}} } {kube-controller-manager Update v1 2022-11-26 05:05:24 +0000 UTC FieldsV1 {"f:metadata":{"f:annotations":{"f:node.alpha.kubernetes.io/ttl":{}}},"f:spec":{"f:podCIDR":{},"f:podCIDRs":{".":{},"v:\"10.64.3.0/24\"":{}}}} } {kube-controller-manager Update v1 2022-11-26 05:17:38 +0000 UTC FieldsV1 {"f:status":{"f:conditions":{"k:{\"type\":\"NetworkUnavailable\"}":{"f:lastHeartbeatTime":{},"f:lastTransitionTime":{},"f:message":{},"f:reason":{},"f:status":{}}}}} status} {node-problem-detector Update v1 2022-11-26 05:20:28 +0000 UTC FieldsV1 {"f:status":{"f:conditions":{"k:{\"type\":\"CorruptDockerOverlay2\"}":{".":{},"f:lastHeartbeatTime":{},"f:lastTransitionTime":{},"f:message":{},"f:reason":{},"f:status":{},"f:type":{}},"k:{\"type\":\"FrequentContainerdRestart\"}":{".":{},"f:lastHeartbeatTime":{},"f:lastTransitionTime":{},"f:message":{},"f:reason":{},"f:status":{},"f:type":{}},"k:{\"type\":\"FrequentDockerRestart\"}":{".":{},"f:lastHeartbeatTime":{},"f:lastTransitionTime":{},"f:message":{},"f:reason":{},"f:status":{},"f:type":{}},"k:{\"type\":\"FrequentKubeletRestart\"}":{".":{},"f:lastHeartbeatTime":{},"f:lastTransitionTime":{},"f:message":{},"f:reason":{},"f:status":{},"f:type":{}},"k:{\"type\":\"FrequentUnregisterNetDevice\"}":{".":{},"f:lastHeartbeatTime":{},"f:lastTransitionTime":{},"f:message":{},"f:reason":{},"f:status":{},"f:type":{}},"k:{\"type\":\"KernelDeadlock\"}":{".":{},"f:lastHeartbeatTime":{},"f:lastTransitionTime":{},"f:message":{},"f:reason":{},"f:status":{},"f:type":{}},"k:{\"type\":\"ReadonlyFilesystem\"}":{".":{},"f:lastHeartbeatTime":{},"f:lastTransitionTime":{},"f:message":{},"f:reason":{},"f:status":{},"f:type":{}}}}} status} {kubelet Update v1 2022-11-26 05:20:53 +0000 UTC FieldsV1 {"f:metadata":{"f:annotations":{"f:csi.volume.kubernetes.io/nodeid":{}},"f:labels":{"f:topology.hostpath.csi/node":{}}},"f:status":{"f:conditions":{"k:{\"type\":\"DiskPressure\"}":{"f:lastHeartbeatTime":{}},"k:{\"type\":\"MemoryPressure\"}":{"f:lastHeartbeatTime":{}},"k:{\"type\":\"PIDPressure\"}":{"f:lastHeartbeatTime":{}},"k:{\"type\":\"Ready\"}":{"f:lastHeartbeatTime":{},"f:lastTransitionTime":{},"f:message":{},"f:reason":{},"f:status":{}}},"f:images":{}}} status}]},Spec:NodeSpec{PodCIDR:10.64.3.0/24,DoNotUseExternalID:,ProviderID:gce://k8s-jkns-e2e-gce-ci-reboot/us-west1-b/bootstrap-e2e-minion-group-vhdj,Unschedulable:false,Taints:[]Taint{},ConfigSource:nil,PodCIDRs:[10.64.3.0/24],},Status:NodeStatus{Capacity:ResourceList{attachable-volumes-gce-pd: {{127 0} {<nil>} 127 DecimalSI},cpu: {{2 0} {<nil>} 2 DecimalSI},ephemeral-storage: {{101203873792 0} {<nil>} 98831908Ki BinarySI},hugepages-1Gi: {{0 0} {<nil>} 0 DecimalSI},hugepages-2Mi: {{0 0} {<nil>} 0 DecimalSI},memory: {{7815438336 0} {<nil>} BinarySI},pods: {{110 0} {<nil>} 110 DecimalSI},},Allocatable:ResourceList{attachable-volumes-gce-pd: {{127 0} {<nil>} 127 DecimalSI},cpu: {{2 0} {<nil>} 2 DecimalSI},ephemeral-storage: {{91083486262 0} {<nil>} 91083486262 DecimalSI},hugepages-1Gi: {{0 0} {<nil>} 0 DecimalSI},hugepages-2Mi: {{0 0} {<nil>} 0 DecimalSI},memory: {{7553294336 0} {<nil>} BinarySI},pods: {{110 0} {<nil>} 110 DecimalSI},},Phase:,Conditions:[]NodeCondition{NodeCondition{Type:FrequentContainerdRestart,Status:False,LastHeartbeatTime:2022-11-26 05:20:28 +0000 UTC,LastTransitionTime:2022-11-26 05:05:26 +0000 UTC,Reason:NoFrequentContainerdRestart,Message:containerd is functioning properly,},NodeCondition{Type:KernelDeadlock,Status:False,LastHeartbeatTime:2022-11-26 05:20:28 +0000 UTC,LastTransitionTime:2022-11-26 05:05:26 +0000 UTC,Reason:KernelHasNoDeadlock,Message:kernel has no deadlock,},NodeCondition{Type:ReadonlyFilesystem,Status:False,LastHeartbeatTime:2022-11-26 05:20:28 +0000 UTC,LastTransitionTime:2022-11-26 05:05:26 +0000 UTC,Reason:FilesystemIsNotReadOnly,Message:Filesystem is not read-only,},NodeCondition{Type:CorruptDockerOverlay2,Status:False,LastHeartbeatTime:2022-11-26 05:20:28 +0000 UTC,LastTransitionTime:2022-11-26 05:05:26 +0000 UTC,Reason:NoCorruptDockerOverlay2,Message:docker overlay2 is functioning properly,},NodeCondition{Type:FrequentUnregisterNetDevice,Status:False,LastHeartbeatTime:2022-11-26 05:20:28 +0000 UTC,LastTransitionTime:2022-11-26 05:05:26 +0000 UTC,Reason:NoFrequentUnregisterNetDevice,Message:node is functioning properly,},NodeCondition{Type:FrequentKubeletRestart,Status:False,LastHeartbeatTime:2022-11-26 05:20:28 +0000 UTC,LastTransitionTime:2022-11-26 05:05:26 +0000 UTC,Reason:NoFrequentKubeletRestart,Message:kubelet is functioning properly,},NodeCondition{Type:FrequentDockerRestart,Status:False,LastHeartbeatTime:2022-11-26 05:20:28 +0000 UTC,LastTransitionTime:2022-11-26 05:05:26 +0000 UTC,Reason:NoFrequentDockerRestart,Message:docker is functioning properly,},NodeCondition{Type:NetworkUnavailable,Status:False,LastHeartbeatTime:2022-11-26 05:05:39 +0000 UTC,LastTransitionTime:2022-11-26 05:05:39 +0000 UTC,Reason:RouteCreated,Message:RouteController created a route,},NodeCondition{Type:MemoryPressure,Status:False,LastHeartbeatTime:2022-11-26 05:17:38 +0000 UTC,LastTransitionTime:2022-11-26 05:05:23 +0000 UTC,Reason:KubeletHasSufficientMemory,Message:kubelet has sufficient memory available,},NodeCondition{Type:DiskPressure,Status:False,LastHeartbeatTime:2022-11-26 05:17:38 +0000 UTC,LastTransitionTime:2022-11-26 05:05:23 +0000 UTC,Reason:KubeletHasNoDiskPressure,Message:kubelet has no disk pressure,},NodeCondition{Type:PIDPressure,Status:False,LastHeartbeatTime:2022-11-26 05:17:38 +0000 UTC,LastTransitionTime:2022-11-26 05:05:23 +0000 UTC,Reason:KubeletHasSufficientPID,Message:kubelet has sufficient PID available,},NodeCondition{Type:Ready,Status:True,LastHeartbeatTime:2022-11-26 05:17:38 +0000 UTC,LastTransitionTime:2022-11-26 05:05:24 +0000 UTC,Reason:KubeletReady,Message:kubelet is posting ready status. AppArmor enabled,},},Addresses:[]NodeAddress{NodeAddress{Type:InternalIP,Address:10.138.0.3,},NodeAddress{Type:ExternalIP,Address:34.83.94.215,},NodeAddress{Type:InternalDNS,Address:bootstrap-e2e-minion-group-vhdj.c.k8s-jkns-e2e-gce-ci-reboot.internal,},NodeAddress{Type:Hostname,Address:bootstrap-e2e-minion-group-vhdj.c.k8s-jkns-e2e-gce-ci-reboot.internal,},},DaemonEndpoints:NodeDaemonEndpoints{KubeletEndpoint:DaemonEndpoint{Port:10250,},},NodeInfo:NodeSystemInfo{MachineID:31b0953ad3e740aeb7fc1a89edb5435a,SystemUUID:31b0953a-d3e7-40ae-b7fc-1a89edb5435a,BootID:14c054d0-bda9-4ddf-b9f6-e1742be33d25,KernelVersion:5.10.123+,OSImage:Container-Optimized OS from Google,ContainerRuntimeVersion:containerd://1.7.0-beta.0-149-gd06318622,KubeletVersion:v1.27.0-alpha.0.50+70617042976dc1,KubeProxyVersion:v1.27.0-alpha.0.50+70617042976dc1,OperatingSystem:linux,Architecture:amd64,},Images:[]ContainerImage{ContainerImage{Names:[registry.k8s.io/e2e-test-images/jessie-dnsutils@sha256:24aaf2626d6b27864c29de2097e8bbb840b3a414271bf7c8995e431e47d8408e registry.k8s.io/e2e-test-images/jessie-dnsutils:1.7],SizeBytes:112030336,},ContainerImage{Names:[registry.k8s.io/sig-storage/nfs-provisioner@sha256:e943bb77c7df05ebdc8c7888b2db289b13bf9f012d6a3a5a74f14d4d5743d439 registry.k8s.io/sig-storage/nfs-provisioner:v3.0.1],SizeBytes:90632047,},ContainerImage{Names:[registry.k8s.io/kube-proxy-amd64:v1.27.0-alpha.0.50_70617042976dc1],SizeBytes:67201736,},ContainerImage{Names:[registry.k8s.io/e2e-test-images/agnhost@sha256:16bbf38c463a4223d8cfe4da12bc61010b082a79b4bb003e2d3ba3ece5dd5f9e registry.k8s.io/e2e-test-images/agnhost:2.43],SizeBytes:51706353,},ContainerImage{Names:[gke.gcr.io/prometheus-to-sd@sha256:e739643c3939ba0b161425f45a1989eedfc4a3b166db9a7100863296b4c70510 gke.gcr.io/prometheus-to-sd:v0.11.1-gke.1],SizeBytes:48742566,},ContainerImage{Names:[registry.k8s.io/e2e-test-images/httpd@sha256:148b022f5c5da426fc2f3c14b5c0867e58ef05961510c84749ac1fddcb0fef22 registry.k8s.io/e2e-test-images/httpd:2.4.38-4],SizeBytes:40764257,},ContainerImage{Names:[registry.k8s.io/sig-storage/csi-provisioner@sha256:ee3b525d5b89db99da3b8eb521d9cd90cb6e9ef0fbb651e98bb37be78d36b5b8 registry.k8s.io/sig-storage/csi-provisioner:v3.3.0],SizeBytes:25491225,},ContainerImage{Names:[registry.k8s.io/sig-storage/csi-resizer@sha256:425d8f1b769398127767b06ed97ce62578a3179bcb99809ce93a1649e025ffe7 registry.k8s.io/sig-storage/csi-resizer:v1.6.0],SizeBytes:24148884,},ContainerImage{Names:[registry.k8s.io/sig-storage/csi-snapshotter@sha256:291334908ddf71a4661fd7f6d9d97274de8a5378a2b6fdfeb2ce73414a34f82f registry.k8s.io/sig-storage/csi-snapshotter:v6.1.0],SizeBytes:23881995,},ContainerImage{Names:[registry.k8s.io/sig-storage/csi-attacher@sha256:9a685020911e2725ad019dbce6e4a5ab93d51e3d4557f115e64343345e05781b registry.k8s.io/sig-storage/csi-attacher:v4.0.0],SizeBytes:23847201,},ContainerImage{Names:[registry.k8s.io/sig-storage/hostpathplugin@sha256:92257881c1d6493cf18299a24af42330f891166560047902b8d431fb66b01af5 registry.k8s.io/sig-storage/hostpathplugin:v1.9.0],SizeBytes:18758628,},ContainerImage{Names:[registry.k8s.io/sig-storage/csi-node-driver-registrar@sha256:0103eee7c35e3e0b5cd8cdca9850dc71c793cdeb6669d8be7a89440da2d06ae4 registry.k8s.io/sig-storage/csi-node-driver-registrar:v2.5.1],SizeBytes:9133109,},ContainerImage{Names:[registry.k8s.io/sig-storage/livenessprobe@sha256:933940f13b3ea0abc62e656c1aa5c5b47c04b15d71250413a6b821bd0c58b94e registry.k8s.io/sig-storage/livenessprobe:v2.7.0],SizeBytes:8688564,},ContainerImage{Names:[registry.k8s.io/kas-network-proxy/proxy-agent@sha256:48f2a4ec3e10553a81b8dd1c6fa5fe4bcc9617f78e71c1ca89c6921335e2d7da registry.k8s.io/kas-network-proxy/proxy-agent:v0.0.33],SizeBytes:8512162,},ContainerImage{Names:[registry.k8s.io/metadata-proxy@sha256:e914645f22e946bce5165737e1b244e0a296ad1f0f81a9531adc57af2780978a registry.k8s.io/metadata-proxy:v0.1.12],SizeBytes:5301657,},ContainerImage{Names:[registry.k8s.io/e2e-test-images/busybox@sha256:2e0f836850e09b8b7cc937681d6194537a09fbd5f6b9e08f4d646a85128e8937 registry.k8s.io/e2e-test-images/busybox:1.29-4],SizeBytes:731990,},ContainerImage{Names:[registry.k8s.io/pause@sha256:9001185023633d17a2f98ff69b6ff2615b8ea02a825adffa40422f51dfdcde9d registry.k8s.io/pause:3.8],SizeBytes:311286,},},VolumesInUse:[],VolumesAttached:[]AttachedVolume{},Config:nil,},} Nov 26 05:20:59.080: INFO: Logging kubelet events for node bootstrap-e2e-minion-group-vhdj Nov 26 05:20:59.128: INFO: Logging pods the kubelet thinks is on node bootstrap-e2e-minion-group-vhdj Nov 26 05:20:59.190: INFO: pod-subpath-test-dynamicpv-4nfp started at 2022-11-26 05:11:17 +0000 UTC (1+2 container statuses recorded) Nov 26 05:20:59.190: INFO: Init container init-volume-dynamicpv-4nfp ready: true, restart count 0 Nov 26 05:20:59.190: INFO: Container test-container-subpath-dynamicpv-4nfp ready: false, restart count 2 Nov 26 05:20:59.190: INFO: Container test-container-volume-dynamicpv-4nfp ready: false, restart count 2 Nov 26 05:20:59.190: INFO: pod-secrets-1f7e1a8b-4bbc-46b8-bd2c-b9a20c65228a started at 2022-11-26 05:16:00 +0000 UTC (0+1 container statuses recorded) Nov 26 05:20:59.190: INFO: Container creates-volume-test ready: false, restart count 0 Nov 26 05:20:59.190: INFO: csi-hostpathplugin-0 started at 2022-11-26 05:16:06 +0000 UTC (0+7 container statuses recorded) Nov 26 05:20:59.190: INFO: Container csi-attacher ready: true, restart count 1 Nov 26 05:20:59.190: INFO: Container csi-provisioner ready: true, restart count 1 Nov 26 05:20:59.190: INFO: Container csi-resizer ready: true, restart count 1 Nov 26 05:20:59.190: INFO: Container csi-snapshotter ready: true, restart count 1 Nov 26 05:20:59.190: INFO: Container hostpath ready: true, restart count 1 Nov 26 05:20:59.190: INFO: Container liveness-probe ready: true, restart count 1 Nov 26 05:20:59.190: INFO: Container node-driver-registrar ready: true, restart count 1 Nov 26 05:20:59.190: INFO: metadata-proxy-v0.1-wn5l5 started at 2022-11-26 05:05:24 +0000 UTC (0+2 container statuses recorded) Nov 26 05:20:59.190: INFO: Container metadata-proxy ready: true, restart count 0 Nov 26 05:20:59.190: INFO: Container prometheus-to-sd-exporter ready: true, restart count 0 Nov 26 05:20:59.190: INFO: csi-hostpathplugin-0 started at 2022-11-26 05:10:50 +0000 UTC (0+7 container statuses recorded) Nov 26 05:20:59.190: INFO: Container csi-attacher ready: true, restart count 4 Nov 26 05:20:59.190: INFO: Container csi-provisioner ready: true, restart count 4 Nov 26 05:20:59.190: INFO: Container csi-resizer ready: true, restart count 4 Nov 26 05:20:59.190: INFO: Container csi-snapshotter ready: true, restart count 4 Nov 26 05:20:59.190: INFO: Container hostpath ready: true, restart count 4 Nov 26 05:20:59.190: INFO: Container liveness-probe ready: true, restart count 4 Nov 26 05:20:59.190: INFO: Container node-driver-registrar ready: true, restart count 4 Nov 26 05:20:59.190: INFO: test-container-pod started at 2022-11-26 05:12:10 +0000 UTC (0+1 container statuses recorded) Nov 26 05:20:59.190: INFO: Container webserver ready: true, restart count 2 Nov 26 05:20:59.190: INFO: forbid-27823995-pbjc6 started at 2022-11-26 05:15:58 +0000 UTC (0+1 container statuses recorded) Nov 26 05:20:59.190: INFO: Container c ready: true, restart count 1 Nov 26 05:20:59.190: INFO: csi-hostpathplugin-0 started at 2022-11-26 05:13:07 +0000 UTC (0+7 container statuses recorded) Nov 26 05:20:59.190: INFO: Container csi-attacher ready: true, restart count 1 Nov 26 05:20:59.190: INFO: Container csi-provisioner ready: true, restart count 1 Nov 26 05:20:59.190: INFO: Container csi-resizer ready: true, restart count 1 Nov 26 05:20:59.190: INFO: Container csi-snapshotter ready: true, restart count 1 Nov 26 05:20:59.190: INFO: Container hostpath ready: true, restart count 1 Nov 26 05:20:59.190: INFO: Container liveness-probe ready: true, restart count 1 Nov 26 05:20:59.190: INFO: Container node-driver-registrar ready: true, restart count 1 Nov 26 05:20:59.190: INFO: csi-hostpathplugin-0 started at 2022-11-26 05:16:00 +0000 UTC (0+7 container statuses recorded) Nov 26 05:20:59.190: INFO: Container csi-attacher ready: true, restart count 4 Nov 26 05:20:59.190: INFO: Container csi-provisioner ready: true, restart count 4 Nov 26 05:20:59.190: INFO: Container csi-resizer ready: true, restart count 4 Nov 26 05:20:59.190: INFO: Container csi-snapshotter ready: true, restart count 4 Nov 26 05:20:59.190: INFO: Container hostpath ready: true, restart count 4 Nov 26 05:20:59.190: INFO: Container liveness-probe ready: true, restart count 4 Nov 26 05:20:59.190: INFO: Container node-driver-registrar ready: true, restart count 4 Nov 26 05:20:59.190: INFO: konnectivity-agent-nt4fd started at 2022-11-26 05:05:39 +0000 UTC (0+1 container statuses recorded) Nov 26 05:20:59.190: INFO: Container konnectivity-agent ready: true, restart count 6 Nov 26 05:20:59.190: INFO: failure-2 started at 2022-11-26 05:16:32 +0000 UTC (0+1 container statuses recorded) Nov 26 05:20:59.190: INFO: Container failure-2 ready: true, restart count 1 Nov 26 05:20:59.190: INFO: csi-hostpathplugin-0 started at 2022-11-26 05:16:44 +0000 UTC (0+7 container statuses recorded) Nov 26 05:20:59.190: INFO: Container csi-attacher ready: true, restart count 1 Nov 26 05:20:59.190: INFO: Container csi-provisioner ready: true, restart count 1 Nov 26 05:20:59.190: INFO: Container csi-resizer ready: true, restart count 1 Nov 26 05:20:59.190: INFO: Container csi-snapshotter ready: true, restart count 1 Nov 26 05:20:59.190: INFO: Container hostpath ready: true, restart count 1 Nov 26 05:20:59.190: INFO: Container liveness-probe ready: true, restart count 1 Nov 26 05:20:59.190: INFO: Container node-driver-registrar ready: true, restart count 1 Nov 26 05:20:59.190: INFO: csi-hostpathplugin-0 started at 2022-11-26 05:16:12 +0000 UTC (0+7 container statuses recorded) Nov 26 05:20:59.190: INFO: Container csi-attacher ready: true, restart count 1 Nov 26 05:20:59.190: INFO: Container csi-provisioner ready: true, restart count 1 Nov 26 05:20:59.190: INFO: Container csi-resizer ready: true, restart count 1 Nov 26 05:20:59.190: INFO: Container csi-snapshotter ready: true, restart count 1 Nov 26 05:20:59.190: INFO: Container hostpath ready: true, restart count 1 Nov 26 05:20:59.190: INFO: Container liveness-probe ready: true, restart count 1 Nov 26 05:20:59.190: INFO: Container node-driver-registrar ready: true, restart count 1 Nov 26 05:20:59.190: INFO: netserver-2 started at 2022-11-26 05:16:25 +0000 UTC (0+1 container statuses recorded) Nov 26 05:20:59.190: INFO: Container webserver ready: false, restart count 5 Nov 26 05:20:59.190: INFO: mutability-test-55dcq started at 2022-11-26 05:13:49 +0000 UTC (0+1 container statuses recorded) Nov 26 05:20:59.190: INFO: Container netexec ready: true, restart count 3 Nov 26 05:20:59.190: INFO: netserver-2 started at 2022-11-26 05:09:51 +0000 UTC (0+1 container statuses recorded) Nov 26 05:20:59.190: INFO: Container webserver ready: true, restart count 3 Nov 26 05:20:59.190: INFO: back-off-cap started at 2022-11-26 05:07:40 +0000 UTC (0+1 container statuses recorded) Nov 26 05:20:59.190: INFO: Container back-off-cap ready: false, restart count 7 Nov 26 05:20:59.190: INFO: netserver-2 started at 2022-11-26 05:11:06 +0000 UTC (0+1 container statuses recorded) Nov 26 05:20:59.190: INFO: Container webserver ready: true, restart count 1 Nov 26 05:20:59.190: INFO: netserver-2 started at 2022-11-26 05:11:29 +0000 UTC (0+1 container statuses recorded) Nov 26 05:20:59.190: INFO: Container webserver ready: false, restart count 3 Nov 26 05:20:59.190: INFO: pod-9fe7e984-9a6a-4bb0-9c65-0628a6a2f792 started at 2022-11-26 05:08:13 +0000 UTC (0+1 container statuses recorded) Nov 26 05:20:59.190: INFO: Container write-pod ready: false, restart count 0 Nov 26 05:20:59.190: INFO: external-local-nodeport-k2k2v started at 2022-11-26 05:16:03 +0000 UTC (0+1 container statuses recorded) Nov 26 05:20:59.190: INFO: Container netexec ready: true, restart count 2 Nov 26 05:20:59.190: INFO: kube-proxy-bootstrap-e2e-minion-group-vhdj started at 2022-11-26 05:05:23 +0000 UTC (0+1 container statuses recorded) Nov 26 05:20:59.190: INFO: Container kube-proxy ready: false, restart count 6 Nov 26 05:20:59.950: INFO: Latency metrics for node bootstrap-e2e-minion-group-vhdj [DeferCleanup (Each)] [sig-network] LoadBalancers ESIPP [Slow] tear down framework | framework.go:193 STEP: Destroying namespace "esipp-5928" for this suite. 11/26/22 05:20:59.95
Filter through log files | View test history on testgrid
go run hack/e2e.go -v --test --test_args='--ginkgo.focus=Kubernetes\se2e\ssuite\s\[It\]\s\[sig\-network\]\sLoadBalancers\sESIPP\s\[Slow\]\sshould\swork\sfrom\spods$'
test/e2e/network/loadbalancer.go:1444 k8s.io/kubernetes/test/e2e/network.glob..func20.6() test/e2e/network/loadbalancer.go:1444 +0x3c6from junit_01.xml
[BeforeEach] [sig-network] LoadBalancers ESIPP [Slow] set up framework | framework.go:178 STEP: Creating a kubernetes client 11/26/22 05:14:00.139 Nov 26 05:14:00.139: INFO: >>> kubeConfig: /workspace/.kube/config STEP: Building a namespace api object, basename esipp 11/26/22 05:14:00.141 Nov 26 05:14:00.180: INFO: Unexpected error while creating namespace: Post "https://35.227.179.144/api/v1/namespaces": dial tcp 35.227.179.144:443: connect: connection refused Nov 26 05:14:02.221: INFO: Unexpected error while creating namespace: Post "https://35.227.179.144/api/v1/namespaces": dial tcp 35.227.179.144:443: connect: connection refused Nov 26 05:14:04.220: INFO: Unexpected error while creating namespace: Post "https://35.227.179.144/api/v1/namespaces": dial tcp 35.227.179.144:443: connect: connection refused Nov 26 05:14:06.221: INFO: Unexpected error while creating namespace: Post "https://35.227.179.144/api/v1/namespaces": dial tcp 35.227.179.144:443: connect: connection refused Nov 26 05:14:08.220: INFO: Unexpected error while creating namespace: Post "https://35.227.179.144/api/v1/namespaces": dial tcp 35.227.179.144:443: connect: connection refused Nov 26 05:14:10.221: INFO: Unexpected error while creating namespace: Post "https://35.227.179.144/api/v1/namespaces": dial tcp 35.227.179.144:443: connect: connection refused Nov 26 05:14:12.221: INFO: Unexpected error while creating namespace: Post "https://35.227.179.144/api/v1/namespaces": dial tcp 35.227.179.144:443: connect: connection refused Nov 26 05:14:14.220: INFO: Unexpected error while creating namespace: Post "https://35.227.179.144/api/v1/namespaces": dial tcp 35.227.179.144:443: connect: connection refused Nov 26 05:14:16.221: INFO: Unexpected error while creating namespace: Post "https://35.227.179.144/api/v1/namespaces": dial tcp 35.227.179.144:443: connect: connection refused Nov 26 05:14:18.220: INFO: Unexpected error while creating namespace: Post "https://35.227.179.144/api/v1/namespaces": dial tcp 35.227.179.144:443: connect: connection refused Nov 26 05:14:20.221: INFO: Unexpected error while creating namespace: Post "https://35.227.179.144/api/v1/namespaces": dial tcp 35.227.179.144:443: connect: connection refused Nov 26 05:14:22.220: INFO: Unexpected error while creating namespace: Post "https://35.227.179.144/api/v1/namespaces": dial tcp 35.227.179.144:443: connect: connection refused Nov 26 05:14:24.220: INFO: Unexpected error while creating namespace: Post "https://35.227.179.144/api/v1/namespaces": dial tcp 35.227.179.144:443: connect: connection refused Nov 26 05:14:26.220: INFO: Unexpected error while creating namespace: Post "https://35.227.179.144/api/v1/namespaces": dial tcp 35.227.179.144:443: connect: connection refused Nov 26 05:14:28.220: INFO: Unexpected error while creating namespace: Post "https://35.227.179.144/api/v1/namespaces": dial tcp 35.227.179.144:443: connect: connection refused Nov 26 05:14:40.220: INFO: Unexpected error while creating namespace: Post "https://35.227.179.144/api/v1/namespaces": net/http: TLS handshake timeout STEP: Waiting for a default service account to be provisioned in namespace 11/26/22 05:15:58.4 STEP: Waiting for kube-root-ca.crt to be provisioned in namespace 11/26/22 05:15:58.492 [BeforeEach] [sig-network] LoadBalancers ESIPP [Slow] test/e2e/framework/metrics/init/init.go:31 [BeforeEach] [sig-network] LoadBalancers ESIPP [Slow] test/e2e/network/loadbalancer.go:1250 [It] should work from pods test/e2e/network/loadbalancer.go:1422 STEP: creating a service esipp-9139/external-local-pods with type=LoadBalancer 11/26/22 05:15:58.77 STEP: setting ExternalTrafficPolicy=Local 11/26/22 05:15:58.77 STEP: waiting for loadbalancer for service esipp-9139/external-local-pods 11/26/22 05:15:58.859 Nov 26 05:15:58.859: INFO: Waiting up to 15m0s for service "external-local-pods" to have a LoadBalancer ------------------------------ Progress Report for Ginkgo Process #2 Automatically polling progress: [sig-network] LoadBalancers ESIPP [Slow] should work from pods (Spec Runtime: 6m58.631s) test/e2e/network/loadbalancer.go:1422 In [It] (Node Runtime: 5m0s) test/e2e/network/loadbalancer.go:1422 At [By Step] waiting for loadbalancer for service esipp-9139/external-local-pods (Step Runtime: 4m59.911s) test/e2e/framework/service/jig.go:260 Spec Goroutine goroutine 2032 [select] k8s.io/kubernetes/vendor/k8s.io/apimachinery/pkg/util/wait.WaitForWithContext({0x7fe0bc8, 0xc000136000}, 0xc00378c240, 0x2fdb16a?) vendor/k8s.io/apimachinery/pkg/util/wait/wait.go:660 k8s.io/kubernetes/vendor/k8s.io/apimachinery/pkg/util/wait.poll({0x7fe0bc8, 0xc000136000}, 0xf0?, 0x2fd9d05?, 0x20?) vendor/k8s.io/apimachinery/pkg/util/wait/wait.go:596 k8s.io/kubernetes/vendor/k8s.io/apimachinery/pkg/util/wait.PollImmediateWithContext({0x7fe0bc8, 0xc000136000}, 0xc0027d8ae0?, 0xc00413ba40?, 0x262a967?) vendor/k8s.io/apimachinery/pkg/util/wait/wait.go:528 k8s.io/kubernetes/vendor/k8s.io/apimachinery/pkg/util/wait.PollImmediate(0xc00187cc70?, 0x7fa7740?, 0xc00021cb40?) vendor/k8s.io/apimachinery/pkg/util/wait/wait.go:514 > k8s.io/kubernetes/test/e2e/framework/service.(*TestJig).waitForCondition(0xc0045ff5e0, 0x4?, {0x7600fe2, 0x14}, 0x7895b68) test/e2e/framework/service/jig.go:631 > k8s.io/kubernetes/test/e2e/framework/service.(*TestJig).WaitForLoadBalancer(0xc0045ff5e0, 0x43?) test/e2e/framework/service/jig.go:582 > k8s.io/kubernetes/test/e2e/framework/service.(*TestJig).CreateLoadBalancerService(0xc0045ff5e0, 0x6aba880?, 0xc00413bcf0) test/e2e/framework/service/jig.go:261 > k8s.io/kubernetes/test/e2e/framework/service.(*TestJig).CreateOnlyLocalLoadBalancerService(0xc0045ff5e0, 0xc003fe7380?, 0x1, 0xa?) test/e2e/framework/service/jig.go:222 > k8s.io/kubernetes/test/e2e/network.glob..func20.6() test/e2e/network/loadbalancer.go:1428 k8s.io/kubernetes/vendor/github.com/onsi/ginkgo/v2/internal.extractBodyFunction.func3({0x2d591ce, 0xc00412a900}) vendor/github.com/onsi/ginkgo/v2/internal/node.go:449 k8s.io/kubernetes/vendor/github.com/onsi/ginkgo/v2/internal.(*Suite).runNode.func2() vendor/github.com/onsi/ginkgo/v2/internal/suite.go:750 k8s.io/kubernetes/vendor/github.com/onsi/ginkgo/v2/internal.(*Suite).runNode vendor/github.com/onsi/ginkgo/v2/internal/suite.go:738 ------------------------------ ------------------------------ Progress Report for Ginkgo Process #2 Automatically polling progress: [sig-network] LoadBalancers ESIPP [Slow] should work from pods (Spec Runtime: 7m18.633s) test/e2e/network/loadbalancer.go:1422 In [It] (Node Runtime: 5m20.002s) test/e2e/network/loadbalancer.go:1422 At [By Step] waiting for loadbalancer for service esipp-9139/external-local-pods (Step Runtime: 5m19.912s) test/e2e/framework/service/jig.go:260 Spec Goroutine goroutine 2032 [select] k8s.io/kubernetes/vendor/k8s.io/apimachinery/pkg/util/wait.WaitForWithContext({0x7fe0bc8, 0xc000136000}, 0xc00378c240, 0x2fdb16a?) vendor/k8s.io/apimachinery/pkg/util/wait/wait.go:660 k8s.io/kubernetes/vendor/k8s.io/apimachinery/pkg/util/wait.poll({0x7fe0bc8, 0xc000136000}, 0xf0?, 0x2fd9d05?, 0x20?) vendor/k8s.io/apimachinery/pkg/util/wait/wait.go:596 k8s.io/kubernetes/vendor/k8s.io/apimachinery/pkg/util/wait.PollImmediateWithContext({0x7fe0bc8, 0xc000136000}, 0xc0027d8ae0?, 0xc00413ba40?, 0x262a967?) vendor/k8s.io/apimachinery/pkg/util/wait/wait.go:528 k8s.io/kubernetes/vendor/k8s.io/apimachinery/pkg/util/wait.PollImmediate(0xc00187cc70?, 0x7fa7740?, 0xc00021cb40?) vendor/k8s.io/apimachinery/pkg/util/wait/wait.go:514 > k8s.io/kubernetes/test/e2e/framework/service.(*TestJig).waitForCondition(0xc0045ff5e0, 0x4?, {0x7600fe2, 0x14}, 0x7895b68) test/e2e/framework/service/jig.go:631 > k8s.io/kubernetes/test/e2e/framework/service.(*TestJig).WaitForLoadBalancer(0xc0045ff5e0, 0x43?) test/e2e/framework/service/jig.go:582 > k8s.io/kubernetes/test/e2e/framework/service.(*TestJig).CreateLoadBalancerService(0xc0045ff5e0, 0x6aba880?, 0xc00413bcf0) test/e2e/framework/service/jig.go:261 > k8s.io/kubernetes/test/e2e/framework/service.(*TestJig).CreateOnlyLocalLoadBalancerService(0xc0045ff5e0, 0xc003fe7380?, 0x1, 0xa?) test/e2e/framework/service/jig.go:222 > k8s.io/kubernetes/test/e2e/network.glob..func20.6() test/e2e/network/loadbalancer.go:1428 k8s.io/kubernetes/vendor/github.com/onsi/ginkgo/v2/internal.extractBodyFunction.func3({0x2d591ce, 0xc00412a900}) vendor/github.com/onsi/ginkgo/v2/internal/node.go:449 k8s.io/kubernetes/vendor/github.com/onsi/ginkgo/v2/internal.(*Suite).runNode.func2() vendor/github.com/onsi/ginkgo/v2/internal/suite.go:750 k8s.io/kubernetes/vendor/github.com/onsi/ginkgo/v2/internal.(*Suite).runNode vendor/github.com/onsi/ginkgo/v2/internal/suite.go:738 ------------------------------ ------------------------------ Progress Report for Ginkgo Process #2 Automatically polling progress: [sig-network] LoadBalancers ESIPP [Slow] should work from pods (Spec Runtime: 7m38.634s) test/e2e/network/loadbalancer.go:1422 In [It] (Node Runtime: 5m40.004s) test/e2e/network/loadbalancer.go:1422 At [By Step] waiting for loadbalancer for service esipp-9139/external-local-pods (Step Runtime: 5m39.914s) test/e2e/framework/service/jig.go:260 Spec Goroutine goroutine 2032 [select] k8s.io/kubernetes/vendor/k8s.io/apimachinery/pkg/util/wait.WaitForWithContext({0x7fe0bc8, 0xc000136000}, 0xc00378c240, 0x2fdb16a?) vendor/k8s.io/apimachinery/pkg/util/wait/wait.go:660 k8s.io/kubernetes/vendor/k8s.io/apimachinery/pkg/util/wait.poll({0x7fe0bc8, 0xc000136000}, 0xf0?, 0x2fd9d05?, 0x20?) vendor/k8s.io/apimachinery/pkg/util/wait/wait.go:596 k8s.io/kubernetes/vendor/k8s.io/apimachinery/pkg/util/wait.PollImmediateWithContext({0x7fe0bc8, 0xc000136000}, 0xc0027d8ae0?, 0xc00413ba40?, 0x262a967?) vendor/k8s.io/apimachinery/pkg/util/wait/wait.go:528 k8s.io/kubernetes/vendor/k8s.io/apimachinery/pkg/util/wait.PollImmediate(0xc00187cc70?, 0x7fa7740?, 0xc00021cb40?) vendor/k8s.io/apimachinery/pkg/util/wait/wait.go:514 > k8s.io/kubernetes/test/e2e/framework/service.(*TestJig).waitForCondition(0xc0045ff5e0, 0x4?, {0x7600fe2, 0x14}, 0x7895b68) test/e2e/framework/service/jig.go:631 > k8s.io/kubernetes/test/e2e/framework/service.(*TestJig).WaitForLoadBalancer(0xc0045ff5e0, 0x43?) test/e2e/framework/service/jig.go:582 > k8s.io/kubernetes/test/e2e/framework/service.(*TestJig).CreateLoadBalancerService(0xc0045ff5e0, 0x6aba880?, 0xc00413bcf0) test/e2e/framework/service/jig.go:261 > k8s.io/kubernetes/test/e2e/framework/service.(*TestJig).CreateOnlyLocalLoadBalancerService(0xc0045ff5e0, 0xc003fe7380?, 0x1, 0xa?) test/e2e/framework/service/jig.go:222 > k8s.io/kubernetes/test/e2e/network.glob..func20.6() test/e2e/network/loadbalancer.go:1428 k8s.io/kubernetes/vendor/github.com/onsi/ginkgo/v2/internal.extractBodyFunction.func3({0x2d591ce, 0xc00412a900}) vendor/github.com/onsi/ginkgo/v2/internal/node.go:449 k8s.io/kubernetes/vendor/github.com/onsi/ginkgo/v2/internal.(*Suite).runNode.func2() vendor/github.com/onsi/ginkgo/v2/internal/suite.go:750 k8s.io/kubernetes/vendor/github.com/onsi/ginkgo/v2/internal.(*Suite).runNode vendor/github.com/onsi/ginkgo/v2/internal/suite.go:738 ------------------------------ ------------------------------ Progress Report for Ginkgo Process #2 Automatically polling progress: [sig-network] LoadBalancers ESIPP [Slow] should work from pods (Spec Runtime: 7m58.637s) test/e2e/network/loadbalancer.go:1422 In [It] (Node Runtime: 6m0.006s) test/e2e/network/loadbalancer.go:1422 At [By Step] waiting for loadbalancer for service esipp-9139/external-local-pods (Step Runtime: 5m59.916s) test/e2e/framework/service/jig.go:260 Spec Goroutine goroutine 2032 [select] k8s.io/kubernetes/vendor/k8s.io/apimachinery/pkg/util/wait.WaitForWithContext({0x7fe0bc8, 0xc000136000}, 0xc00378c240, 0x2fdb16a?) vendor/k8s.io/apimachinery/pkg/util/wait/wait.go:660 k8s.io/kubernetes/vendor/k8s.io/apimachinery/pkg/util/wait.poll({0x7fe0bc8, 0xc000136000}, 0xf0?, 0x2fd9d05?, 0x20?) vendor/k8s.io/apimachinery/pkg/util/wait/wait.go:596 k8s.io/kubernetes/vendor/k8s.io/apimachinery/pkg/util/wait.PollImmediateWithContext({0x7fe0bc8, 0xc000136000}, 0xc0027d8ae0?, 0xc00413ba40?, 0x262a967?) vendor/k8s.io/apimachinery/pkg/util/wait/wait.go:528 k8s.io/kubernetes/vendor/k8s.io/apimachinery/pkg/util/wait.PollImmediate(0xc00187cc70?, 0x7fa7740?, 0xc00021cb40?) vendor/k8s.io/apimachinery/pkg/util/wait/wait.go:514 > k8s.io/kubernetes/test/e2e/framework/service.(*TestJig).waitForCondition(0xc0045ff5e0, 0x4?, {0x7600fe2, 0x14}, 0x7895b68) test/e2e/framework/service/jig.go:631 > k8s.io/kubernetes/test/e2e/framework/service.(*TestJig).WaitForLoadBalancer(0xc0045ff5e0, 0x43?) test/e2e/framework/service/jig.go:582 > k8s.io/kubernetes/test/e2e/framework/service.(*TestJig).CreateLoadBalancerService(0xc0045ff5e0, 0x6aba880?, 0xc00413bcf0) test/e2e/framework/service/jig.go:261 > k8s.io/kubernetes/test/e2e/framework/service.(*TestJig).CreateOnlyLocalLoadBalancerService(0xc0045ff5e0, 0xc003fe7380?, 0x1, 0xa?) test/e2e/framework/service/jig.go:222 > k8s.io/kubernetes/test/e2e/network.glob..func20.6() test/e2e/network/loadbalancer.go:1428 k8s.io/kubernetes/vendor/github.com/onsi/ginkgo/v2/internal.extractBodyFunction.func3({0x2d591ce, 0xc00412a900}) vendor/github.com/onsi/ginkgo/v2/internal/node.go:449 k8s.io/kubernetes/vendor/github.com/onsi/ginkgo/v2/internal.(*Suite).runNode.func2() vendor/github.com/onsi/ginkgo/v2/internal/suite.go:750 k8s.io/kubernetes/vendor/github.com/onsi/ginkgo/v2/internal.(*Suite).runNode vendor/github.com/onsi/ginkgo/v2/internal/suite.go:738 ------------------------------ ------------------------------ Progress Report for Ginkgo Process #2 Automatically polling progress: [sig-network] LoadBalancers ESIPP [Slow] should work from pods (Spec Runtime: 8m18.638s) test/e2e/network/loadbalancer.go:1422 In [It] (Node Runtime: 6m20.007s) test/e2e/network/loadbalancer.go:1422 At [By Step] waiting for loadbalancer for service esipp-9139/external-local-pods (Step Runtime: 6m19.918s) test/e2e/framework/service/jig.go:260 Spec Goroutine goroutine 2032 [select] k8s.io/kubernetes/vendor/k8s.io/apimachinery/pkg/util/wait.WaitForWithContext({0x7fe0bc8, 0xc000136000}, 0xc00378c240, 0x2fdb16a?) vendor/k8s.io/apimachinery/pkg/util/wait/wait.go:660 k8s.io/kubernetes/vendor/k8s.io/apimachinery/pkg/util/wait.poll({0x7fe0bc8, 0xc000136000}, 0xf0?, 0x2fd9d05?, 0x20?) vendor/k8s.io/apimachinery/pkg/util/wait/wait.go:596 k8s.io/kubernetes/vendor/k8s.io/apimachinery/pkg/util/wait.PollImmediateWithContext({0x7fe0bc8, 0xc000136000}, 0xc0027d8ae0?, 0xc00413ba40?, 0x262a967?) vendor/k8s.io/apimachinery/pkg/util/wait/wait.go:528 k8s.io/kubernetes/vendor/k8s.io/apimachinery/pkg/util/wait.PollImmediate(0xc00187cc70?, 0x7fa7740?, 0xc00021cb40?) vendor/k8s.io/apimachinery/pkg/util/wait/wait.go:514 > k8s.io/kubernetes/test/e2e/framework/service.(*TestJig).waitForCondition(0xc0045ff5e0, 0x4?, {0x7600fe2, 0x14}, 0x7895b68) test/e2e/framework/service/jig.go:631 > k8s.io/kubernetes/test/e2e/framework/service.(*TestJig).WaitForLoadBalancer(0xc0045ff5e0, 0x43?) test/e2e/framework/service/jig.go:582 > k8s.io/kubernetes/test/e2e/framework/service.(*TestJig).CreateLoadBalancerService(0xc0045ff5e0, 0x6aba880?, 0xc00413bcf0) test/e2e/framework/service/jig.go:261 > k8s.io/kubernetes/test/e2e/framework/service.(*TestJig).CreateOnlyLocalLoadBalancerService(0xc0045ff5e0, 0xc003fe7380?, 0x1, 0xa?) test/e2e/framework/service/jig.go:222 > k8s.io/kubernetes/test/e2e/network.glob..func20.6() test/e2e/network/loadbalancer.go:1428 k8s.io/kubernetes/vendor/github.com/onsi/ginkgo/v2/internal.extractBodyFunction.func3({0x2d591ce, 0xc00412a900}) vendor/github.com/onsi/ginkgo/v2/internal/node.go:449 k8s.io/kubernetes/vendor/github.com/onsi/ginkgo/v2/internal.(*Suite).runNode.func2() vendor/github.com/onsi/ginkgo/v2/internal/suite.go:750 k8s.io/kubernetes/vendor/github.com/onsi/ginkgo/v2/internal.(*Suite).runNode vendor/github.com/onsi/ginkgo/v2/internal/suite.go:738 ------------------------------ ------------------------------ Progress Report for Ginkgo Process #2 Automatically polling progress: [sig-network] LoadBalancers ESIPP [Slow] should work from pods (Spec Runtime: 8m38.64s) test/e2e/network/loadbalancer.go:1422 In [It] (Node Runtime: 6m40.009s) test/e2e/network/loadbalancer.go:1422 At [By Step] waiting for loadbalancer for service esipp-9139/external-local-pods (Step Runtime: 6m39.92s) test/e2e/framework/service/jig.go:260 Spec Goroutine goroutine 2032 [select] k8s.io/kubernetes/vendor/k8s.io/apimachinery/pkg/util/wait.WaitForWithContext({0x7fe0bc8, 0xc000136000}, 0xc00378c240, 0x2fdb16a?) vendor/k8s.io/apimachinery/pkg/util/wait/wait.go:660 k8s.io/kubernetes/vendor/k8s.io/apimachinery/pkg/util/wait.poll({0x7fe0bc8, 0xc000136000}, 0xf0?, 0x2fd9d05?, 0x20?) vendor/k8s.io/apimachinery/pkg/util/wait/wait.go:596 k8s.io/kubernetes/vendor/k8s.io/apimachinery/pkg/util/wait.PollImmediateWithContext({0x7fe0bc8, 0xc000136000}, 0xc0027d8ae0?, 0xc00413ba40?, 0x262a967?) vendor/k8s.io/apimachinery/pkg/util/wait/wait.go:528 k8s.io/kubernetes/vendor/k8s.io/apimachinery/pkg/util/wait.PollImmediate(0xc00187cc70?, 0x7fa7740?, 0xc00021cb40?) vendor/k8s.io/apimachinery/pkg/util/wait/wait.go:514 > k8s.io/kubernetes/test/e2e/framework/service.(*TestJig).waitForCondition(0xc0045ff5e0, 0x4?, {0x7600fe2, 0x14}, 0x7895b68) test/e2e/framework/service/jig.go:631 > k8s.io/kubernetes/test/e2e/framework/service.(*TestJig).WaitForLoadBalancer(0xc0045ff5e0, 0x43?) test/e2e/framework/service/jig.go:582 > k8s.io/kubernetes/test/e2e/framework/service.(*TestJig).CreateLoadBalancerService(0xc0045ff5e0, 0x6aba880?, 0xc00413bcf0) test/e2e/framework/service/jig.go:261 > k8s.io/kubernetes/test/e2e/framework/service.(*TestJig).CreateOnlyLocalLoadBalancerService(0xc0045ff5e0, 0xc003fe7380?, 0x1, 0xa?) test/e2e/framework/service/jig.go:222 > k8s.io/kubernetes/test/e2e/network.glob..func20.6() test/e2e/network/loadbalancer.go:1428 k8s.io/kubernetes/vendor/github.com/onsi/ginkgo/v2/internal.extractBodyFunction.func3({0x2d591ce, 0xc00412a900}) vendor/github.com/onsi/ginkgo/v2/internal/node.go:449 k8s.io/kubernetes/vendor/github.com/onsi/ginkgo/v2/internal.(*Suite).runNode.func2() vendor/github.com/onsi/ginkgo/v2/internal/suite.go:750 k8s.io/kubernetes/vendor/github.com/onsi/ginkgo/v2/internal.(*Suite).runNode vendor/github.com/onsi/ginkgo/v2/internal/suite.go:738 ------------------------------ ------------------------------ Progress Report for Ginkgo Process #2 Automatically polling progress: [sig-network] LoadBalancers ESIPP [Slow] should work from pods (Spec Runtime: 8m58.642s) test/e2e/network/loadbalancer.go:1422 In [It] (Node Runtime: 7m0.011s) test/e2e/network/loadbalancer.go:1422 At [By Step] waiting for loadbalancer for service esipp-9139/external-local-pods (Step Runtime: 6m59.921s) test/e2e/framework/service/jig.go:260 Spec Goroutine goroutine 2032 [select] k8s.io/kubernetes/vendor/k8s.io/apimachinery/pkg/util/wait.WaitForWithContext({0x7fe0bc8, 0xc000136000}, 0xc00378c240, 0x2fdb16a?) vendor/k8s.io/apimachinery/pkg/util/wait/wait.go:660 k8s.io/kubernetes/vendor/k8s.io/apimachinery/pkg/util/wait.poll({0x7fe0bc8, 0xc000136000}, 0xf0?, 0x2fd9d05?, 0x20?) vendor/k8s.io/apimachinery/pkg/util/wait/wait.go:596 k8s.io/kubernetes/vendor/k8s.io/apimachinery/pkg/util/wait.PollImmediateWithContext({0x7fe0bc8, 0xc000136000}, 0xc0027d8ae0?, 0xc00413ba40?, 0x262a967?) vendor/k8s.io/apimachinery/pkg/util/wait/wait.go:528 k8s.io/kubernetes/vendor/k8s.io/apimachinery/pkg/util/wait.PollImmediate(0xc00187cc70?, 0x7fa7740?, 0xc00021cb40?) vendor/k8s.io/apimachinery/pkg/util/wait/wait.go:514 > k8s.io/kubernetes/test/e2e/framework/service.(*TestJig).waitForCondition(0xc0045ff5e0, 0x4?, {0x7600fe2, 0x14}, 0x7895b68) test/e2e/framework/service/jig.go:631 > k8s.io/kubernetes/test/e2e/framework/service.(*TestJig).WaitForLoadBalancer(0xc0045ff5e0, 0x43?) test/e2e/framework/service/jig.go:582 > k8s.io/kubernetes/test/e2e/framework/service.(*TestJig).CreateLoadBalancerService(0xc0045ff5e0, 0x6aba880?, 0xc00413bcf0) test/e2e/framework/service/jig.go:261 > k8s.io/kubernetes/test/e2e/framework/service.(*TestJig).CreateOnlyLocalLoadBalancerService(0xc0045ff5e0, 0xc003fe7380?, 0x1, 0xa?) test/e2e/framework/service/jig.go:222 > k8s.io/kubernetes/test/e2e/network.glob..func20.6() test/e2e/network/loadbalancer.go:1428 k8s.io/kubernetes/vendor/github.com/onsi/ginkgo/v2/internal.extractBodyFunction.func3({0x2d591ce, 0xc00412a900}) vendor/github.com/onsi/ginkgo/v2/internal/node.go:449 k8s.io/kubernetes/vendor/github.com/onsi/ginkgo/v2/internal.(*Suite).runNode.func2() vendor/github.com/onsi/ginkgo/v2/internal/suite.go:750 k8s.io/kubernetes/vendor/github.com/onsi/ginkgo/v2/internal.(*Suite).runNode vendor/github.com/onsi/ginkgo/v2/internal/suite.go:738 ------------------------------ ------------------------------ Progress Report for Ginkgo Process #2 Automatically polling progress: [sig-network] LoadBalancers ESIPP [Slow] should work from pods (Spec Runtime: 9m18.644s) test/e2e/network/loadbalancer.go:1422 In [It] (Node Runtime: 7m20.013s) test/e2e/network/loadbalancer.go:1422 At [By Step] waiting for loadbalancer for service esipp-9139/external-local-pods (Step Runtime: 7m19.923s) test/e2e/framework/service/jig.go:260 Spec Goroutine goroutine 2032 [select] k8s.io/kubernetes/vendor/k8s.io/apimachinery/pkg/util/wait.WaitForWithContext({0x7fe0bc8, 0xc000136000}, 0xc00378c240, 0x2fdb16a?) vendor/k8s.io/apimachinery/pkg/util/wait/wait.go:660 k8s.io/kubernetes/vendor/k8s.io/apimachinery/pkg/util/wait.poll({0x7fe0bc8, 0xc000136000}, 0xf0?, 0x2fd9d05?, 0x20?) vendor/k8s.io/apimachinery/pkg/util/wait/wait.go:596 k8s.io/kubernetes/vendor/k8s.io/apimachinery/pkg/util/wait.PollImmediateWithContext({0x7fe0bc8, 0xc000136000}, 0xc0027d8ae0?, 0xc00413ba40?, 0x262a967?) vendor/k8s.io/apimachinery/pkg/util/wait/wait.go:528 k8s.io/kubernetes/vendor/k8s.io/apimachinery/pkg/util/wait.PollImmediate(0xc00187cc70?, 0x7fa7740?, 0xc00021cb40?) vendor/k8s.io/apimachinery/pkg/util/wait/wait.go:514 > k8s.io/kubernetes/test/e2e/framework/service.(*TestJig).waitForCondition(0xc0045ff5e0, 0x4?, {0x7600fe2, 0x14}, 0x7895b68) test/e2e/framework/service/jig.go:631 > k8s.io/kubernetes/test/e2e/framework/service.(*TestJig).WaitForLoadBalancer(0xc0045ff5e0, 0x43?) test/e2e/framework/service/jig.go:582 > k8s.io/kubernetes/test/e2e/framework/service.(*TestJig).CreateLoadBalancerService(0xc0045ff5e0, 0x6aba880?, 0xc00413bcf0) test/e2e/framework/service/jig.go:261 > k8s.io/kubernetes/test/e2e/framework/service.(*TestJig).CreateOnlyLocalLoadBalancerService(0xc0045ff5e0, 0xc003fe7380?, 0x1, 0xa?) test/e2e/framework/service/jig.go:222 > k8s.io/kubernetes/test/e2e/network.glob..func20.6() test/e2e/network/loadbalancer.go:1428 k8s.io/kubernetes/vendor/github.com/onsi/ginkgo/v2/internal.extractBodyFunction.func3({0x2d591ce, 0xc00412a900}) vendor/github.com/onsi/ginkgo/v2/internal/node.go:449 k8s.io/kubernetes/vendor/github.com/onsi/ginkgo/v2/internal.(*Suite).runNode.func2() vendor/github.com/onsi/ginkgo/v2/internal/suite.go:750 k8s.io/kubernetes/vendor/github.com/onsi/ginkgo/v2/internal.(*Suite).runNode vendor/github.com/onsi/ginkgo/v2/internal/suite.go:738 ------------------------------ ------------------------------ Progress Report for Ginkgo Process #2 Automatically polling progress: [sig-network] LoadBalancers ESIPP [Slow] should work from pods (Spec Runtime: 9m38.646s) test/e2e/network/loadbalancer.go:1422 In [It] (Node Runtime: 7m40.015s) test/e2e/network/loadbalancer.go:1422 At [By Step] waiting for loadbalancer for service esipp-9139/external-local-pods (Step Runtime: 7m39.926s) test/e2e/framework/service/jig.go:260 Spec Goroutine goroutine 2032 [select] k8s.io/kubernetes/vendor/k8s.io/apimachinery/pkg/util/wait.WaitForWithContext({0x7fe0bc8, 0xc000136000}, 0xc00378c240, 0x2fdb16a?) vendor/k8s.io/apimachinery/pkg/util/wait/wait.go:660 k8s.io/kubernetes/vendor/k8s.io/apimachinery/pkg/util/wait.poll({0x7fe0bc8, 0xc000136000}, 0xf0?, 0x2fd9d05?, 0x20?) vendor/k8s.io/apimachinery/pkg/util/wait/wait.go:596 k8s.io/kubernetes/vendor/k8s.io/apimachinery/pkg/util/wait.PollImmediateWithContext({0x7fe0bc8, 0xc000136000}, 0xc0027d8ae0?, 0xc00413ba40?, 0x262a967?) vendor/k8s.io/apimachinery/pkg/util/wait/wait.go:528 k8s.io/kubernetes/vendor/k8s.io/apimachinery/pkg/util/wait.PollImmediate(0xc00187cc70?, 0x7fa7740?, 0xc00021cb40?) vendor/k8s.io/apimachinery/pkg/util/wait/wait.go:514 > k8s.io/kubernetes/test/e2e/framework/service.(*TestJig).waitForCondition(0xc0045ff5e0, 0x4?, {0x7600fe2, 0x14}, 0x7895b68) test/e2e/framework/service/jig.go:631 > k8s.io/kubernetes/test/e2e/framework/service.(*TestJig).WaitForLoadBalancer(0xc0045ff5e0, 0x43?) test/e2e/framework/service/jig.go:582 > k8s.io/kubernetes/test/e2e/framework/service.(*TestJig).CreateLoadBalancerService(0xc0045ff5e0, 0x6aba880?, 0xc00413bcf0) test/e2e/framework/service/jig.go:261 > k8s.io/kubernetes/test/e2e/framework/service.(*TestJig).CreateOnlyLocalLoadBalancerService(0xc0045ff5e0, 0xc003fe7380?, 0x1, 0xa?) test/e2e/framework/service/jig.go:222 > k8s.io/kubernetes/test/e2e/network.glob..func20.6() test/e2e/network/loadbalancer.go:1428 k8s.io/kubernetes/vendor/github.com/onsi/ginkgo/v2/internal.extractBodyFunction.func3({0x2d591ce, 0xc00412a900}) vendor/github.com/onsi/ginkgo/v2/internal/node.go:449 k8s.io/kubernetes/vendor/github.com/onsi/ginkgo/v2/internal.(*Suite).runNode.func2() vendor/github.com/onsi/ginkgo/v2/internal/suite.go:750 k8s.io/kubernetes/vendor/github.com/onsi/ginkgo/v2/internal.(*Suite).runNode vendor/github.com/onsi/ginkgo/v2/internal/suite.go:738 ------------------------------ STEP: creating a pod to be part of the service external-local-pods 11/26/22 05:23:40.978 Nov 26 05:23:41.073: INFO: Waiting up to 2m0s for 1 pods to be created Nov 26 05:23:41.156: INFO: Found all 1 pods Nov 26 05:23:41.156: INFO: Waiting up to 2m0s for 1 pods to be running and ready: [external-local-pods-4s55w] Nov 26 05:23:41.156: INFO: Waiting up to 2m0s for pod "external-local-pods-4s55w" in namespace "esipp-9139" to be "running and ready" Nov 26 05:23:41.225: INFO: Pod "external-local-pods-4s55w": Phase="Pending", Reason="", readiness=false. Elapsed: 69.283328ms Nov 26 05:23:41.225: INFO: Error evaluating pod condition running and ready: want pod 'external-local-pods-4s55w' on 'bootstrap-e2e-minion-group-vhdj' to be 'Running' but was 'Pending' Nov 26 05:23:43.421: INFO: Pod "external-local-pods-4s55w": Phase="Pending", Reason="", readiness=false. Elapsed: 2.265017366s Nov 26 05:23:43.421: INFO: Error evaluating pod condition running and ready: want pod 'external-local-pods-4s55w' on 'bootstrap-e2e-minion-group-vhdj' to be 'Running' but was 'Pending' Nov 26 05:23:45.301: INFO: Pod "external-local-pods-4s55w": Phase="Pending", Reason="", readiness=false. Elapsed: 4.145542315s Nov 26 05:23:45.301: INFO: Error evaluating pod condition running and ready: want pod 'external-local-pods-4s55w' on 'bootstrap-e2e-minion-group-vhdj' to be 'Running' but was 'Pending' Nov 26 05:23:47.376: INFO: Pod "external-local-pods-4s55w": Phase="Running", Reason="", readiness=true. Elapsed: 6.219911521s Nov 26 05:23:47.376: INFO: Pod "external-local-pods-4s55w" satisfied condition "running and ready" Nov 26 05:23:47.376: INFO: Wanted all 1 pods to be running and ready. Result: true. Pods: [external-local-pods-4s55w] STEP: waiting for loadbalancer for service esipp-9139/external-local-pods 11/26/22 05:23:47.376 Nov 26 05:23:47.376: INFO: Waiting up to 15m0s for service "external-local-pods" to have a LoadBalancer STEP: Creating pause pod deployment to make sure, pausePods are in desired state 11/26/22 05:23:47.417 Nov 26 05:23:47.516: INFO: deployment status: v1.DeploymentStatus{ObservedGeneration:0, Replicas:0, UpdatedReplicas:0, ReadyReplicas:0, AvailableReplicas:0, UnavailableReplicas:0, Conditions:[]v1.DeploymentCondition(nil), CollisionCount:(*int32)(nil)} Nov 26 05:23:49.558: INFO: deployment status: v1.DeploymentStatus{ObservedGeneration:0, Replicas:0, UpdatedReplicas:0, ReadyReplicas:0, AvailableReplicas:0, UnavailableReplicas:0, Conditions:[]v1.DeploymentCondition(nil), CollisionCount:(*int32)(nil)} Nov 26 05:23:51.557: INFO: deployment status: v1.DeploymentStatus{ObservedGeneration:0, Replicas:0, UpdatedReplicas:0, ReadyReplicas:0, AvailableReplicas:0, UnavailableReplicas:0, Conditions:[]v1.DeploymentCondition(nil), CollisionCount:(*int32)(nil)} Nov 26 05:23:53.560: INFO: deployment status: v1.DeploymentStatus{ObservedGeneration:0, Replicas:0, UpdatedReplicas:0, ReadyReplicas:0, AvailableReplicas:0, UnavailableReplicas:0, Conditions:[]v1.DeploymentCondition(nil), CollisionCount:(*int32)(nil)} Nov 26 05:23:55.558: INFO: deployment status: v1.DeploymentStatus{ObservedGeneration:0, Replicas:0, UpdatedReplicas:0, ReadyReplicas:0, AvailableReplicas:0, UnavailableReplicas:0, Conditions:[]v1.DeploymentCondition(nil), CollisionCount:(*int32)(nil)} Nov 26 05:23:57.558: INFO: deployment status: v1.DeploymentStatus{ObservedGeneration:0, Replicas:0, UpdatedReplicas:0, ReadyReplicas:0, AvailableReplicas:0, UnavailableReplicas:0, Conditions:[]v1.DeploymentCondition(nil), CollisionCount:(*int32)(nil)} ------------------------------ Progress Report for Ginkgo Process #2 Automatically polling progress: [sig-network] LoadBalancers ESIPP [Slow] should work from pods (Spec Runtime: 9m58.648s) test/e2e/network/loadbalancer.go:1422 In [It] (Node Runtime: 8m0.018s) test/e2e/network/loadbalancer.go:1422 At [By Step] Creating pause pod deployment to make sure, pausePods are in desired state (Step Runtime: 11.37s) test/e2e/network/loadbalancer.go:1442 Spec Goroutine goroutine 2032 [select] k8s.io/kubernetes/vendor/k8s.io/apimachinery/pkg/util/wait.WaitForWithContext({0x7fe0bc8, 0xc000136000}, 0xc00557e5d0, 0x2fdb16a?) vendor/k8s.io/apimachinery/pkg/util/wait/wait.go:660 k8s.io/kubernetes/vendor/k8s.io/apimachinery/pkg/util/wait.poll({0x7fe0bc8, 0xc000136000}, 0x8?, 0x2fd9d05?, 0x40?) vendor/k8s.io/apimachinery/pkg/util/wait/wait.go:596 k8s.io/kubernetes/vendor/k8s.io/apimachinery/pkg/util/wait.PollImmediateWithContext({0x7fe0bc8, 0xc000136000}, 0x0?, 0xc00413bc58?, 0x262a967?) vendor/k8s.io/apimachinery/pkg/util/wait/wait.go:528 k8s.io/kubernetes/vendor/k8s.io/apimachinery/pkg/util/wait.PollImmediate(0x0?, 0x0?, 0x0?) vendor/k8s.io/apimachinery/pkg/util/wait/wait.go:514 k8s.io/kubernetes/test/utils.waitForDeploymentCompleteMaybeCheckRolling({0x801de88?, 0xc003fe7380}, 0xc000cc3900, 0x0, 0x78959b8, 0xa?, 0x1?) test/utils/deployment.go:82 k8s.io/kubernetes/test/utils.WaitForDeploymentComplete(...) test/utils/deployment.go:201 k8s.io/kubernetes/test/e2e/framework/deployment.WaitForDeploymentComplete(...) test/e2e/framework/deployment/wait.go:46 > k8s.io/kubernetes/test/e2e/network.glob..func20.6() test/e2e/network/loadbalancer.go:1444 k8s.io/kubernetes/vendor/github.com/onsi/ginkgo/v2/internal.extractBodyFunction.func3({0x2d591ce, 0xc00412a900}) vendor/github.com/onsi/ginkgo/v2/internal/node.go:449 k8s.io/kubernetes/vendor/github.com/onsi/ginkgo/v2/internal.(*Suite).runNode.func2() vendor/github.com/onsi/ginkgo/v2/internal/suite.go:750 k8s.io/kubernetes/vendor/github.com/onsi/ginkgo/v2/internal.(*Suite).runNode vendor/github.com/onsi/ginkgo/v2/internal/suite.go:738 ------------------------------ Nov 26 05:23:59.558: INFO: deployment status: v1.DeploymentStatus{ObservedGeneration:0, Replicas:0, UpdatedReplicas:0, ReadyReplicas:0, AvailableReplicas:0, UnavailableReplicas:0, Conditions:[]v1.DeploymentCondition(nil), CollisionCount:(*int32)(nil)} Nov 26 05:24:01.558: INFO: deployment status: v1.DeploymentStatus{ObservedGeneration:0, Replicas:0, UpdatedReplicas:0, ReadyReplicas:0, AvailableReplicas:0, UnavailableReplicas:0, Conditions:[]v1.DeploymentCondition(nil), CollisionCount:(*int32)(nil)} Nov 26 05:24:03.570: INFO: deployment status: v1.DeploymentStatus{ObservedGeneration:0, Replicas:0, UpdatedReplicas:0, ReadyReplicas:0, AvailableReplicas:0, UnavailableReplicas:0, Conditions:[]v1.DeploymentCondition(nil), CollisionCount:(*int32)(nil)} Nov 26 05:24:05.557: INFO: deployment status: v1.DeploymentStatus{ObservedGeneration:0, Replicas:0, UpdatedReplicas:0, ReadyReplicas:0, AvailableReplicas:0, UnavailableReplicas:0, Conditions:[]v1.DeploymentCondition(nil), CollisionCount:(*int32)(nil)} Nov 26 05:24:07.558: INFO: deployment status: v1.DeploymentStatus{ObservedGeneration:0, Replicas:0, UpdatedReplicas:0, ReadyReplicas:0, AvailableReplicas:0, UnavailableReplicas:0, Conditions:[]v1.DeploymentCondition(nil), CollisionCount:(*int32)(nil)} Nov 26 05:24:09.558: INFO: deployment status: v1.DeploymentStatus{ObservedGeneration:0, Replicas:0, UpdatedReplicas:0, ReadyReplicas:0, AvailableReplicas:0, UnavailableReplicas:0, Conditions:[]v1.DeploymentCondition(nil), CollisionCount:(*int32)(nil)} Nov 26 05:24:11.559: INFO: deployment status: v1.DeploymentStatus{ObservedGeneration:0, Replicas:0, UpdatedReplicas:0, ReadyReplicas:0, AvailableReplicas:0, UnavailableReplicas:0, Conditions:[]v1.DeploymentCondition(nil), CollisionCount:(*int32)(nil)} Nov 26 05:24:13.560: INFO: deployment status: v1.DeploymentStatus{ObservedGeneration:0, Replicas:0, UpdatedReplicas:0, ReadyReplicas:0, AvailableReplicas:0, UnavailableReplicas:0, Conditions:[]v1.DeploymentCondition(nil), CollisionCount:(*int32)(nil)} Nov 26 05:24:15.561: INFO: deployment status: v1.DeploymentStatus{ObservedGeneration:0, Replicas:0, UpdatedReplicas:0, ReadyReplicas:0, AvailableReplicas:0, UnavailableReplicas:0, Conditions:[]v1.DeploymentCondition(nil), CollisionCount:(*int32)(nil)} Nov 26 05:24:17.557: INFO: deployment status: v1.DeploymentStatus{ObservedGeneration:0, Replicas:0, UpdatedReplicas:0, ReadyReplicas:0, AvailableReplicas:0, UnavailableReplicas:0, Conditions:[]v1.DeploymentCondition(nil), CollisionCount:(*int32)(nil)} ------------------------------ Progress Report for Ginkgo Process #2 Automatically polling progress: [sig-network] LoadBalancers ESIPP [Slow] should work from pods (Spec Runtime: 10m18.651s) test/e2e/network/loadbalancer.go:1422 In [It] (Node Runtime: 8m20.02s) test/e2e/network/loadbalancer.go:1422 At [By Step] Creating pause pod deployment to make sure, pausePods are in desired state (Step Runtime: 31.373s) test/e2e/network/loadbalancer.go:1442 Spec Goroutine goroutine 2032 [select] k8s.io/kubernetes/vendor/k8s.io/apimachinery/pkg/util/wait.WaitForWithContext({0x7fe0bc8, 0xc000136000}, 0xc00557e5d0, 0x2fdb16a?) vendor/k8s.io/apimachinery/pkg/util/wait/wait.go:660 k8s.io/kubernetes/vendor/k8s.io/apimachinery/pkg/util/wait.poll({0x7fe0bc8, 0xc000136000}, 0x8?, 0x2fd9d05?, 0x40?) vendor/k8s.io/apimachinery/pkg/util/wait/wait.go:596 k8s.io/kubernetes/vendor/k8s.io/apimachinery/pkg/util/wait.PollImmediateWithContext({0x7fe0bc8, 0xc000136000}, 0x0?, 0xc00413bc58?, 0x262a967?) vendor/k8s.io/apimachinery/pkg/util/wait/wait.go:528 k8s.io/kubernetes/vendor/k8s.io/apimachinery/pkg/util/wait.PollImmediate(0x0?, 0x0?, 0x0?) vendor/k8s.io/apimachinery/pkg/util/wait/wait.go:514 k8s.io/kubernetes/test/utils.waitForDeploymentCompleteMaybeCheckRolling({0x801de88?, 0xc003fe7380}, 0xc000cc3900, 0x0, 0x78959b8, 0xa?, 0x1?) test/utils/deployment.go:82 k8s.io/kubernetes/test/utils.WaitForDeploymentComplete(...) test/utils/deployment.go:201 k8s.io/kubernetes/test/e2e/framework/deployment.WaitForDeploymentComplete(...) test/e2e/framework/deployment/wait.go:46 > k8s.io/kubernetes/test/e2e/network.glob..func20.6() test/e2e/network/loadbalancer.go:1444 k8s.io/kubernetes/vendor/github.com/onsi/ginkgo/v2/internal.extractBodyFunction.func3({0x2d591ce, 0xc00412a900}) vendor/github.com/onsi/ginkgo/v2/internal/node.go:449 k8s.io/kubernetes/vendor/github.com/onsi/ginkgo/v2/internal.(*Suite).runNode.func2() vendor/github.com/onsi/ginkgo/v2/internal/suite.go:750 k8s.io/kubernetes/vendor/github.com/onsi/ginkgo/v2/internal.(*Suite).runNode vendor/github.com/onsi/ginkgo/v2/internal/suite.go:738 ------------------------------ Nov 26 05:24:19.558: INFO: deployment status: v1.DeploymentStatus{ObservedGeneration:0, Replicas:0, UpdatedReplicas:0, ReadyReplicas:0, AvailableReplicas:0, UnavailableReplicas:0, Conditions:[]v1.DeploymentCondition(nil), CollisionCount:(*int32)(nil)} Nov 26 05:24:21.557: INFO: deployment status: v1.DeploymentStatus{ObservedGeneration:0, Replicas:0, UpdatedReplicas:0, ReadyReplicas:0, AvailableReplicas:0, UnavailableReplicas:0, Conditions:[]v1.DeploymentCondition(nil), CollisionCount:(*int32)(nil)} Nov 26 05:24:23.559: INFO: deployment status: v1.DeploymentStatus{ObservedGeneration:0, Replicas:0, UpdatedReplicas:0, ReadyReplicas:0, AvailableReplicas:0, UnavailableReplicas:0, Conditions:[]v1.DeploymentCondition(nil), CollisionCount:(*int32)(nil)} Nov 26 05:24:25.558: INFO: deployment status: v1.DeploymentStatus{ObservedGeneration:0, Replicas:0, UpdatedReplicas:0, ReadyReplicas:0, AvailableReplicas:0, UnavailableReplicas:0, Conditions:[]v1.DeploymentCondition(nil), CollisionCount:(*int32)(nil)} Nov 26 05:24:27.557: INFO: deployment status: v1.DeploymentStatus{ObservedGeneration:0, Replicas:0, UpdatedReplicas:0, ReadyReplicas:0, AvailableReplicas:0, UnavailableReplicas:0, Conditions:[]v1.DeploymentCondition(nil), CollisionCount:(*int32)(nil)} Nov 26 05:24:29.558: INFO: deployment status: v1.DeploymentStatus{ObservedGeneration:0, Replicas:0, UpdatedReplicas:0, ReadyReplicas:0, AvailableReplicas:0, UnavailableReplicas:0, Conditions:[]v1.DeploymentCondition(nil), CollisionCount:(*int32)(nil)} Nov 26 05:24:31.557: INFO: deployment status: v1.DeploymentStatus{ObservedGeneration:0, Replicas:0, UpdatedReplicas:0, ReadyReplicas:0, AvailableReplicas:0, UnavailableReplicas:0, Conditions:[]v1.DeploymentCondition(nil), CollisionCount:(*int32)(nil)} Nov 26 05:24:33.562: INFO: deployment status: v1.DeploymentStatus{ObservedGeneration:0, Replicas:0, UpdatedReplicas:0, ReadyReplicas:0, AvailableReplicas:0, UnavailableReplicas:0, Conditions:[]v1.DeploymentCondition(nil), CollisionCount:(*int32)(nil)} Nov 26 05:24:35.558: INFO: deployment status: v1.DeploymentStatus{ObservedGeneration:0, Replicas:0, UpdatedReplicas:0, ReadyReplicas:0, AvailableReplicas:0, UnavailableReplicas:0, Conditions:[]v1.DeploymentCondition(nil), CollisionCount:(*int32)(nil)} Nov 26 05:24:37.557: INFO: deployment status: v1.DeploymentStatus{ObservedGeneration:0, Replicas:0, UpdatedReplicas:0, ReadyReplicas:0, AvailableReplicas:0, UnavailableReplicas:0, Conditions:[]v1.DeploymentCondition(nil), CollisionCount:(*int32)(nil)} ------------------------------ Progress Report for Ginkgo Process #2 Automatically polling progress: [sig-network] LoadBalancers ESIPP [Slow] should work from pods (Spec Runtime: 10m38.652s) test/e2e/network/loadbalancer.go:1422 In [It] (Node Runtime: 8m40.022s) test/e2e/network/loadbalancer.go:1422 At [By Step] Creating pause pod deployment to make sure, pausePods are in desired state (Step Runtime: 51.374s) test/e2e/network/loadbalancer.go:1442 Spec Goroutine goroutine 2032 [select] k8s.io/kubernetes/vendor/k8s.io/apimachinery/pkg/util/wait.WaitForWithContext({0x7fe0bc8, 0xc000136000}, 0xc00557e5d0, 0x2fdb16a?) vendor/k8s.io/apimachinery/pkg/util/wait/wait.go:660 k8s.io/kubernetes/vendor/k8s.io/apimachinery/pkg/util/wait.poll({0x7fe0bc8, 0xc000136000}, 0x8?, 0x2fd9d05?, 0x40?) vendor/k8s.io/apimachinery/pkg/util/wait/wait.go:596 k8s.io/kubernetes/vendor/k8s.io/apimachinery/pkg/util/wait.PollImmediateWithContext({0x7fe0bc8, 0xc000136000}, 0x0?, 0xc00413bc58?, 0x262a967?) vendor/k8s.io/apimachinery/pkg/util/wait/wait.go:528 k8s.io/kubernetes/vendor/k8s.io/apimachinery/pkg/util/wait.PollImmediate(0x0?, 0x0?, 0x0?) vendor/k8s.io/apimachinery/pkg/util/wait/wait.go:514 k8s.io/kubernetes/test/utils.waitForDeploymentCompleteMaybeCheckRolling({0x801de88?, 0xc003fe7380}, 0xc000cc3900, 0x0, 0x78959b8, 0xa?, 0x1?) test/utils/deployment.go:82 k8s.io/kubernetes/test/utils.WaitForDeploymentComplete(...) test/utils/deployment.go:201 k8s.io/kubernetes/test/e2e/framework/deployment.WaitForDeploymentComplete(...) test/e2e/framework/deployment/wait.go:46 > k8s.io/kubernetes/test/e2e/network.glob..func20.6() test/e2e/network/loadbalancer.go:1444 k8s.io/kubernetes/vendor/github.com/onsi/ginkgo/v2/internal.extractBodyFunction.func3({0x2d591ce, 0xc00412a900}) vendor/github.com/onsi/ginkgo/v2/internal/node.go:449 k8s.io/kubernetes/vendor/github.com/onsi/ginkgo/v2/internal.(*Suite).runNode.func2() vendor/github.com/onsi/ginkgo/v2/internal/suite.go:750 k8s.io/kubernetes/vendor/github.com/onsi/ginkgo/v2/internal.(*Suite).runNode vendor/github.com/onsi/ginkgo/v2/internal/suite.go:738 ------------------------------ Nov 26 05:24:39.559: INFO: deployment status: v1.DeploymentStatus{ObservedGeneration:0, Replicas:0, UpdatedReplicas:0, ReadyReplicas:0, AvailableReplicas:0, UnavailableReplicas:0, Conditions:[]v1.DeploymentCondition(nil), CollisionCount:(*int32)(nil)} Nov 26 05:24:41.560: INFO: deployment status: v1.DeploymentStatus{ObservedGeneration:0, Replicas:0, UpdatedReplicas:0, ReadyReplicas:0, AvailableReplicas:0, UnavailableReplicas:0, Conditions:[]v1.DeploymentCondition(nil), CollisionCount:(*int32)(nil)} Nov 26 05:24:43.561: INFO: deployment status: v1.DeploymentStatus{ObservedGeneration:0, Replicas:0, UpdatedReplicas:0, ReadyReplicas:0, AvailableReplicas:0, UnavailableReplicas:0, Conditions:[]v1.DeploymentCondition(nil), CollisionCount:(*int32)(nil)} Nov 26 05:24:45.558: INFO: deployment status: v1.DeploymentStatus{ObservedGeneration:0, Replicas:0, UpdatedReplicas:0, ReadyReplicas:0, AvailableReplicas:0, UnavailableReplicas:0, Conditions:[]v1.DeploymentCondition(nil), CollisionCount:(*int32)(nil)} Nov 26 05:24:47.561: INFO: deployment status: v1.DeploymentStatus{ObservedGeneration:0, Replicas:0, UpdatedReplicas:0, ReadyReplicas:0, AvailableReplicas:0, UnavailableReplicas:0, Conditions:[]v1.DeploymentCondition(nil), CollisionCount:(*int32)(nil)} Nov 26 05:24:49.557: INFO: deployment status: v1.DeploymentStatus{ObservedGeneration:0, Replicas:0, UpdatedReplicas:0, ReadyReplicas:0, AvailableReplicas:0, UnavailableReplicas:0, Conditions:[]v1.DeploymentCondition(nil), CollisionCount:(*int32)(nil)} Nov 26 05:24:51.558: INFO: deployment status: v1.DeploymentStatus{ObservedGeneration:0, Replicas:0, UpdatedReplicas:0, ReadyReplicas:0, AvailableReplicas:0, UnavailableReplicas:0, Conditions:[]v1.DeploymentCondition(nil), CollisionCount:(*int32)(nil)} Nov 26 05:24:53.559: INFO: deployment status: v1.DeploymentStatus{ObservedGeneration:0, Replicas:0, UpdatedReplicas:0, ReadyReplicas:0, AvailableReplicas:0, UnavailableReplicas:0, Conditions:[]v1.DeploymentCondition(nil), CollisionCount:(*int32)(nil)} Nov 26 05:24:55.557: INFO: deployment status: v1.DeploymentStatus{ObservedGeneration:0, Replicas:0, UpdatedReplicas:0, ReadyReplicas:0, AvailableReplicas:0, UnavailableReplicas:0, Conditions:[]v1.DeploymentCondition(nil), CollisionCount:(*int32)(nil)} Nov 26 05:24:57.558: INFO: deployment status: v1.DeploymentStatus{ObservedGeneration:0, Replicas:0, UpdatedReplicas:0, ReadyReplicas:0, AvailableReplicas:0, UnavailableReplicas:0, Conditions:[]v1.DeploymentCondition(nil), CollisionCount:(*int32)(nil)} ------------------------------ Progress Report for Ginkgo Process #2 Automatically polling progress: [sig-network] LoadBalancers ESIPP [Slow] should work from pods (Spec Runtime: 10m58.655s) test/e2e/network/loadbalancer.go:1422 In [It] (Node Runtime: 9m0.024s) test/e2e/network/loadbalancer.go:1422 At [By Step] Creating pause pod deployment to make sure, pausePods are in desired state (Step Runtime: 1m11.377s) test/e2e/network/loadbalancer.go:1442 Spec Goroutine goroutine 2032 [select] k8s.io/kubernetes/vendor/k8s.io/apimachinery/pkg/util/wait.WaitForWithContext({0x7fe0bc8, 0xc000136000}, 0xc00557e5d0, 0x2fdb16a?) vendor/k8s.io/apimachinery/pkg/util/wait/wait.go:660 k8s.io/kubernetes/vendor/k8s.io/apimachinery/pkg/util/wait.poll({0x7fe0bc8, 0xc000136000}, 0x8?, 0x2fd9d05?, 0x40?) vendor/k8s.io/apimachinery/pkg/util/wait/wait.go:596 k8s.io/kubernetes/vendor/k8s.io/apimachinery/pkg/util/wait.PollImmediateWithContext({0x7fe0bc8, 0xc000136000}, 0x0?, 0xc00413bc58?, 0x262a967?) vendor/k8s.io/apimachinery/pkg/util/wait/wait.go:528 k8s.io/kubernetes/vendor/k8s.io/apimachinery/pkg/util/wait.PollImmediate(0x0?, 0x0?, 0x0?) vendor/k8s.io/apimachinery/pkg/util/wait/wait.go:514 k8s.io/kubernetes/test/utils.waitForDeploymentCompleteMaybeCheckRolling({0x801de88?, 0xc003fe7380}, 0xc000cc3900, 0x0, 0x78959b8, 0xa?, 0x1?) test/utils/deployment.go:82 k8s.io/kubernetes/test/utils.WaitForDeploymentComplete(...) test/utils/deployment.go:201 k8s.io/kubernetes/test/e2e/framework/deployment.WaitForDeploymentComplete(...) test/e2e/framework/deployment/wait.go:46 > k8s.io/kubernetes/test/e2e/network.glob..func20.6() test/e2e/network/loadbalancer.go:1444 k8s.io/kubernetes/vendor/github.com/onsi/ginkgo/v2/internal.extractBodyFunction.func3({0x2d591ce, 0xc00412a900}) vendor/github.com/onsi/ginkgo/v2/internal/node.go:449 k8s.io/kubernetes/vendor/github.com/onsi/ginkgo/v2/internal.(*Suite).runNode.func2() vendor/github.com/onsi/ginkgo/v2/internal/suite.go:750 k8s.io/kubernetes/vendor/github.com/onsi/ginkgo/v2/internal.(*Suite).runNode vendor/github.com/onsi/ginkgo/v2/internal/suite.go:738 ------------------------------ Nov 26 05:24:59.558: INFO: deployment status: v1.DeploymentStatus{ObservedGeneration:0, Replicas:0, UpdatedReplicas:0, ReadyReplicas:0, AvailableReplicas:0, UnavailableReplicas:0, Conditions:[]v1.DeploymentCondition(nil), CollisionCount:(*int32)(nil)} Nov 26 05:25:01.558: INFO: deployment status: v1.DeploymentStatus{ObservedGeneration:0, Replicas:0, UpdatedReplicas:0, ReadyReplicas:0, AvailableReplicas:0, UnavailableReplicas:0, Conditions:[]v1.DeploymentCondition(nil), CollisionCount:(*int32)(nil)} Nov 26 05:25:03.565: INFO: deployment status: v1.DeploymentStatus{ObservedGeneration:0, Replicas:0, UpdatedReplicas:0, ReadyReplicas:0, AvailableReplicas:0, UnavailableReplicas:0, Conditions:[]v1.DeploymentCondition(nil), CollisionCount:(*int32)(nil)} Nov 26 05:25:05.559: INFO: deployment status: v1.DeploymentStatus{ObservedGeneration:0, Replicas:0, UpdatedReplicas:0, ReadyReplicas:0, AvailableReplicas:0, UnavailableReplicas:0, Conditions:[]v1.DeploymentCondition(nil), CollisionCount:(*int32)(nil)} Nov 26 05:25:07.558: INFO: deployment status: v1.DeploymentStatus{ObservedGeneration:0, Replicas:0, UpdatedReplicas:0, ReadyReplicas:0, AvailableReplicas:0, UnavailableReplicas:0, Conditions:[]v1.DeploymentCondition(nil), CollisionCount:(*int32)(nil)} Nov 26 05:25:09.557: INFO: deployment status: v1.DeploymentStatus{ObservedGeneration:0, Replicas:0, UpdatedReplicas:0, ReadyReplicas:0, AvailableReplicas:0, UnavailableReplicas:0, Conditions:[]v1.DeploymentCondition(nil), CollisionCount:(*int32)(nil)} Nov 26 05:25:11.558: INFO: deployment status: v1.DeploymentStatus{ObservedGeneration:0, Replicas:0, UpdatedReplicas:0, ReadyReplicas:0, AvailableReplicas:0, UnavailableReplicas:0, Conditions:[]v1.DeploymentCondition(nil), CollisionCount:(*int32)(nil)} Nov 26 05:25:13.559: INFO: deployment status: v1.DeploymentStatus{ObservedGeneration:0, Replicas:0, UpdatedReplicas:0, ReadyReplicas:0, AvailableReplicas:0, UnavailableReplicas:0, Conditions:[]v1.DeploymentCondition(nil), CollisionCount:(*int32)(nil)} Nov 26 05:25:15.558: INFO: deployment status: v1.DeploymentStatus{ObservedGeneration:0, Replicas:0, UpdatedReplicas:0, ReadyReplicas:0, AvailableReplicas:0, UnavailableReplicas:0, Conditions:[]v1.DeploymentCondition(nil), CollisionCount:(*int32)(nil)} Nov 26 05:25:17.557: INFO: deployment status: v1.DeploymentStatus{ObservedGeneration:0, Replicas:0, UpdatedReplicas:0, ReadyReplicas:0, AvailableReplicas:0, UnavailableReplicas:0, Conditions:[]v1.DeploymentCondition(nil), CollisionCount:(*int32)(nil)} ------------------------------ Progress Report for Ginkgo Process #2 Automatically polling progress: [sig-network] LoadBalancers ESIPP [Slow] should work from pods (Spec Runtime: 11m18.657s) test/e2e/network/loadbalancer.go:1422 In [It] (Node Runtime: 9m20.026s) test/e2e/network/loadbalancer.go:1422 At [By Step] Creating pause pod deployment to make sure, pausePods are in desired state (Step Runtime: 1m31.379s) test/e2e/network/loadbalancer.go:1442 Spec Goroutine goroutine 2032 [select] k8s.io/kubernetes/vendor/k8s.io/apimachinery/pkg/util/wait.WaitForWithContext({0x7fe0bc8, 0xc000136000}, 0xc00557e5d0, 0x2fdb16a?) vendor/k8s.io/apimachinery/pkg/util/wait/wait.go:660 k8s.io/kubernetes/vendor/k8s.io/apimachinery/pkg/util/wait.poll({0x7fe0bc8, 0xc000136000}, 0x8?, 0x2fd9d05?, 0x40?) vendor/k8s.io/apimachinery/pkg/util/wait/wait.go:596 k8s.io/kubernetes/vendor/k8s.io/apimachinery/pkg/util/wait.PollImmediateWithContext({0x7fe0bc8, 0xc000136000}, 0x0?, 0xc00413bc58?, 0x262a967?) vendor/k8s.io/apimachinery/pkg/util/wait/wait.go:528 k8s.io/kubernetes/vendor/k8s.io/apimachinery/pkg/util/wait.PollImmediate(0x0?, 0x0?, 0x0?) vendor/k8s.io/apimachinery/pkg/util/wait/wait.go:514 k8s.io/kubernetes/test/utils.waitForDeploymentCompleteMaybeCheckRolling({0x801de88?, 0xc003fe7380}, 0xc000cc3900, 0x0, 0x78959b8, 0xa?, 0x1?) test/utils/deployment.go:82 k8s.io/kubernetes/test/utils.WaitForDeploymentComplete(...) test/utils/deployment.go:201 k8s.io/kubernetes/test/e2e/framework/deployment.WaitForDeploymentComplete(...) test/e2e/framework/deployment/wait.go:46 > k8s.io/kubernetes/test/e2e/network.glob..func20.6() test/e2e/network/loadbalancer.go:1444 k8s.io/kubernetes/vendor/github.com/onsi/ginkgo/v2/internal.extractBodyFunction.func3({0x2d591ce, 0xc00412a900}) vendor/github.com/onsi/ginkgo/v2/internal/node.go:449 k8s.io/kubernetes/vendor/github.com/onsi/ginkgo/v2/internal.(*Suite).runNode.func2() vendor/github.com/onsi/ginkgo/v2/internal/suite.go:750 k8s.io/kubernetes/vendor/github.com/onsi/ginkgo/v2/internal.(*Suite).runNode vendor/github.com/onsi/ginkgo/v2/internal/suite.go:738 ------------------------------ Nov 26 05:25:19.557: INFO: deployment status: v1.DeploymentStatus{ObservedGeneration:0, Replicas:0, UpdatedReplicas:0, ReadyReplicas:0, AvailableReplicas:0, UnavailableReplicas:0, Conditions:[]v1.DeploymentCondition(nil), CollisionCount:(*int32)(nil)} Nov 26 05:25:21.560: INFO: deployment status: v1.DeploymentStatus{ObservedGeneration:0, Replicas:0, UpdatedReplicas:0, ReadyReplicas:0, AvailableReplicas:0, UnavailableReplicas:0, Conditions:[]v1.DeploymentCondition(nil), CollisionCount:(*int32)(nil)} Nov 26 05:25:23.565: INFO: deployment status: v1.DeploymentStatus{ObservedGeneration:0, Replicas:0, UpdatedReplicas:0, ReadyReplicas:0, AvailableReplicas:0, UnavailableReplicas:0, Conditions:[]v1.DeploymentCondition(nil), CollisionCount:(*int32)(nil)} Nov 26 05:25:25.558: INFO: deployment status: v1.DeploymentStatus{ObservedGeneration:0, Replicas:0, UpdatedReplicas:0, ReadyReplicas:0, AvailableReplicas:0, UnavailableReplicas:0, Conditions:[]v1.DeploymentCondition(nil), CollisionCount:(*int32)(nil)} Nov 26 05:25:27.558: INFO: deployment status: v1.DeploymentStatus{ObservedGeneration:0, Replicas:0, UpdatedReplicas:0, ReadyReplicas:0, AvailableReplicas:0, UnavailableReplicas:0, Conditions:[]v1.DeploymentCondition(nil), CollisionCount:(*int32)(nil)} Nov 26 05:25:29.559: INFO: deployment status: v1.DeploymentStatus{ObservedGeneration:0, Replicas:0, UpdatedReplicas:0, ReadyReplicas:0, AvailableReplicas:0, UnavailableReplicas:0, Conditions:[]v1.DeploymentCondition(nil), CollisionCount:(*int32)(nil)} Nov 26 05:25:31.559: INFO: deployment status: v1.DeploymentStatus{ObservedGeneration:0, Replicas:0, UpdatedReplicas:0, ReadyReplicas:0, AvailableReplicas:0, UnavailableReplicas:0, Conditions:[]v1.DeploymentCondition(nil), CollisionCount:(*int32)(nil)} Nov 26 05:25:33.561: INFO: deployment status: v1.DeploymentStatus{ObservedGeneration:0, Replicas:0, UpdatedReplicas:0, ReadyReplicas:0, AvailableReplicas:0, UnavailableReplicas:0, Conditions:[]v1.DeploymentCondition(nil), CollisionCount:(*int32)(nil)} Nov 26 05:25:35.558: INFO: deployment status: v1.DeploymentStatus{ObservedGeneration:0, Replicas:0, UpdatedReplicas:0, ReadyReplicas:0, AvailableReplicas:0, UnavailableReplicas:0, Conditions:[]v1.DeploymentCondition(nil), CollisionCount:(*int32)(nil)} Nov 26 05:25:37.559: INFO: deployment status: v1.DeploymentStatus{ObservedGeneration:0, Replicas:0, UpdatedReplicas:0, ReadyReplicas:0, AvailableReplicas:0, UnavailableReplicas:0, Conditions:[]v1.DeploymentCondition(nil), CollisionCount:(*int32)(nil)} ------------------------------ Progress Report for Ginkgo Process #2 Automatically polling progress: [sig-network] LoadBalancers ESIPP [Slow] should work from pods (Spec Runtime: 11m38.66s) test/e2e/network/loadbalancer.go:1422 In [It] (Node Runtime: 9m40.029s) test/e2e/network/loadbalancer.go:1422 At [By Step] Creating pause pod deployment to make sure, pausePods are in desired state (Step Runtime: 1m51.382s) test/e2e/network/loadbalancer.go:1442 Spec Goroutine goroutine 2032 [select] k8s.io/kubernetes/vendor/k8s.io/apimachinery/pkg/util/wait.WaitForWithContext({0x7fe0bc8, 0xc000136000}, 0xc00557e5d0, 0x2fdb16a?) vendor/k8s.io/apimachinery/pkg/util/wait/wait.go:660 k8s.io/kubernetes/vendor/k8s.io/apimachinery/pkg/util/wait.poll({0x7fe0bc8, 0xc000136000}, 0x8?, 0x2fd9d05?, 0x40?) vendor/k8s.io/apimachinery/pkg/util/wait/wait.go:596 k8s.io/kubernetes/vendor/k8s.io/apimachinery/pkg/util/wait.PollImmediateWithContext({0x7fe0bc8, 0xc000136000}, 0x0?, 0xc00413bc58?, 0x262a967?) vendor/k8s.io/apimachinery/pkg/util/wait/wait.go:528 k8s.io/kubernetes/vendor/k8s.io/apimachinery/pkg/util/wait.PollImmediate(0x0?, 0x0?, 0x0?) vendor/k8s.io/apimachinery/pkg/util/wait/wait.go:514 k8s.io/kubernetes/test/utils.waitForDeploymentCompleteMaybeCheckRolling({0x801de88?, 0xc003fe7380}, 0xc000cc3900, 0x0, 0x78959b8, 0xa?, 0x1?) test/utils/deployment.go:82 k8s.io/kubernetes/test/utils.WaitForDeploymentComplete(...) test/utils/deployment.go:201 k8s.io/kubernetes/test/e2e/framework/deployment.WaitForDeploymentComplete(...) test/e2e/framework/deployment/wait.go:46 > k8s.io/kubernetes/test/e2e/network.glob..func20.6() test/e2e/network/loadbalancer.go:1444 k8s.io/kubernetes/vendor/github.com/onsi/ginkgo/v2/internal.extractBodyFunction.func3({0x2d591ce, 0xc00412a900}) vendor/github.com/onsi/ginkgo/v2/internal/node.go:449 k8s.io/kubernetes/vendor/github.com/onsi/ginkgo/v2/internal.(*Suite).runNode.func2() vendor/github.com/onsi/ginkgo/v2/internal/suite.go:750 k8s.io/kubernetes/vendor/github.com/onsi/ginkgo/v2/internal.(*Suite).runNode vendor/github.com/onsi/ginkgo/v2/internal/suite.go:738 ------------------------------ Nov 26 05:25:39.593: INFO: deployment status: v1.DeploymentStatus{ObservedGeneration:0, Replicas:0, UpdatedReplicas:0, ReadyReplicas:0, AvailableReplicas:0, UnavailableReplicas:0, Conditions:[]v1.DeploymentCondition(nil), CollisionCount:(*int32)(nil)} ------------------------------ Progress Report for Ginkgo Process #2 Automatically polling progress: [sig-network] LoadBalancers ESIPP [Slow] should work from pods (Spec Runtime: 11m58.662s) test/e2e/network/loadbalancer.go:1422 In [It] (Node Runtime: 10m0.031s) test/e2e/network/loadbalancer.go:1422 At [By Step] Creating pause pod deployment to make sure, pausePods are in desired state (Step Runtime: 2m11.384s) test/e2e/network/loadbalancer.go:1442 Spec Goroutine goroutine 2032 [select] k8s.io/kubernetes/vendor/golang.org/x/net/http2.(*ClientConn).RoundTrip(0xc000cec600, 0xc0038a4300) vendor/golang.org/x/net/http2/transport.go:1200 k8s.io/kubernetes/vendor/golang.org/x/net/http2.(*Transport).RoundTripOpt(0xc0040ec400, 0xc0038a4300, {0xe0?}) vendor/golang.org/x/net/http2/transport.go:519 k8s.io/kubernetes/vendor/golang.org/x/net/http2.(*Transport).RoundTrip(...) vendor/golang.org/x/net/http2/transport.go:480 k8s.io/kubernetes/vendor/golang.org/x/net/http2.noDialH2RoundTripper.RoundTrip({0xc0040f6000?}, 0xc0038a4300?) vendor/golang.org/x/net/http2/transport.go:3020 net/http.(*Transport).roundTrip(0xc0040f6000, 0xc0038a4300) /usr/local/go/src/net/http/transport.go:540 net/http.(*Transport).RoundTrip(0x6fe4b20?, 0xc0051f6810?) /usr/local/go/src/net/http/roundtrip.go:17 k8s.io/kubernetes/vendor/k8s.io/client-go/transport.(*bearerAuthRoundTripper).RoundTrip(0xc001dc9530, 0xc0038a4200) vendor/k8s.io/client-go/transport/round_trippers.go:317 k8s.io/kubernetes/vendor/k8s.io/client-go/transport.(*userAgentRoundTripper).RoundTrip(0xc0016b67e0, 0xc0038a4100) vendor/k8s.io/client-go/transport/round_trippers.go:168 net/http.send(0xc0038a4100, {0x7fad100, 0xc0016b67e0}, {0x74d54e0?, 0x1?, 0x0?}) /usr/local/go/src/net/http/client.go:251 net/http.(*Client).send(0xc001dc9560, 0xc0038a4100, {0x7f9d4c43e108?, 0x100?, 0x0?}) /usr/local/go/src/net/http/client.go:175 net/http.(*Client).do(0xc001dc9560, 0xc0038a4100) /usr/local/go/src/net/http/client.go:715 net/http.(*Client).Do(...) /usr/local/go/src/net/http/client.go:581 k8s.io/kubernetes/vendor/k8s.io/client-go/rest.(*Request).request(0xc000ca1d00, {0x7fe0bc8, 0xc000136008}, 0x0?) vendor/k8s.io/client-go/rest/request.go:964 k8s.io/kubernetes/vendor/k8s.io/client-go/rest.(*Request).Do(0xc000ca1d00, {0x7fe0bc8, 0xc000136008}) vendor/k8s.io/client-go/rest/request.go:1005 k8s.io/kubernetes/vendor/k8s.io/client-go/kubernetes/typed/apps/v1.(*deployments).Get(0xc0018642e0, {0x7fe0bc8, 0xc000136008}, {0xc00557e570, 0x14}, {{{0x0, 0x0}, {0x0, 0x0}}, {0x0, ...}}) vendor/k8s.io/client-go/kubernetes/typed/apps/v1/deployment.go:86 k8s.io/kubernetes/test/utils.waitForDeploymentCompleteMaybeCheckRolling.func1() test/utils/deployment.go:84 k8s.io/kubernetes/vendor/k8s.io/apimachinery/pkg/util/wait.ConditionFunc.WithContext.func1({0x2742871, 0x0}) vendor/k8s.io/apimachinery/pkg/util/wait/wait.go:222 k8s.io/kubernetes/vendor/k8s.io/apimachinery/pkg/util/wait.runConditionWithCrashProtectionWithContext({0x7fe0bc8?, 0xc000136000?}, 0xc003fe7380?) vendor/k8s.io/apimachinery/pkg/util/wait/wait.go:235 k8s.io/kubernetes/vendor/k8s.io/apimachinery/pkg/util/wait.WaitForWithContext({0x7fe0bc8, 0xc000136000}, 0xc00557e5d0, 0x2fdb16a?) vendor/k8s.io/apimachinery/pkg/util/wait/wait.go:662 k8s.io/kubernetes/vendor/k8s.io/apimachinery/pkg/util/wait.poll({0x7fe0bc8, 0xc000136000}, 0x8?, 0x2fd9d05?, 0x40?) vendor/k8s.io/apimachinery/pkg/util/wait/wait.go:596 k8s.io/kubernetes/vendor/k8s.io/apimachinery/pkg/util/wait.PollImmediateWithContext({0x7fe0bc8, 0xc000136000}, 0x0?, 0xc00413bc58?, 0x262a967?) vendor/k8s.io/apimachinery/pkg/util/wait/wait.go:528 k8s.io/kubernetes/vendor/k8s.io/apimachinery/pkg/util/wait.PollImmediate(0x0?, 0x0?, 0x0?) vendor/k8s.io/apimachinery/pkg/util/wait/wait.go:514 k8s.io/kubernetes/test/utils.waitForDeploymentCompleteMaybeCheckRolling({0x801de88?, 0xc003fe7380}, 0xc000cc3900, 0x0, 0x78959b8, 0xa?, 0x1?) test/utils/deployment.go:82 k8s.io/kubernetes/test/utils.WaitForDeploymentComplete(...) test/utils/deployment.go:201 k8s.io/kubernetes/test/e2e/framework/deployment.WaitForDeploymentComplete(...) test/e2e/framework/deployment/wait.go:46 > k8s.io/kubernetes/test/e2e/network.glob..func20.6() test/e2e/network/loadbalancer.go:1444 k8s.io/kubernetes/vendor/github.com/onsi/ginkgo/v2/internal.extractBodyFunction.func3({0x2d591ce, 0xc00412a900}) vendor/github.com/onsi/ginkgo/v2/internal/node.go:449 k8s.io/kubernetes/vendor/github.com/onsi/ginkgo/v2/internal.(*Suite).runNode.func2() vendor/github.com/onsi/ginkgo/v2/internal/suite.go:750 k8s.io/kubernetes/vendor/github.com/onsi/ginkgo/v2/internal.(*Suite).runNode vendor/github.com/onsi/ginkgo/v2/internal/suite.go:738 ------------------------------ ------------------------------ Progress Report for Ginkgo Process #2 Automatically polling progress: [sig-network] LoadBalancers ESIPP [Slow] should work from pods (Spec Runtime: 12m18.665s) test/e2e/network/loadbalancer.go:1422 In [It] (Node Runtime: 10m20.034s) test/e2e/network/loadbalancer.go:1422 At [By Step] Creating pause pod deployment to make sure, pausePods are in desired state (Step Runtime: 2m31.387s) test/e2e/network/loadbalancer.go:1442 Spec Goroutine goroutine 2032 [select] k8s.io/kubernetes/vendor/golang.org/x/net/http2.(*ClientConn).RoundTrip(0xc000cec600, 0xc0038a4300) vendor/golang.org/x/net/http2/transport.go:1200 k8s.io/kubernetes/vendor/golang.org/x/net/http2.(*Transport).RoundTripOpt(0xc0040ec400, 0xc0038a4300, {0xe0?}) vendor/golang.org/x/net/http2/transport.go:519 k8s.io/kubernetes/vendor/golang.org/x/net/http2.(*Transport).RoundTrip(...) vendor/golang.org/x/net/http2/transport.go:480 k8s.io/kubernetes/vendor/golang.org/x/net/http2.noDialH2RoundTripper.RoundTrip({0xc0040f6000?}, 0xc0038a4300?) vendor/golang.org/x/net/http2/transport.go:3020 net/http.(*Transport).roundTrip(0xc0040f6000, 0xc0038a4300) /usr/local/go/src/net/http/transport.go:540 net/http.(*Transport).RoundTrip(0x6fe4b20?, 0xc0051f6810?) /usr/local/go/src/net/http/roundtrip.go:17 k8s.io/kubernetes/vendor/k8s.io/client-go/transport.(*bearerAuthRoundTripper).RoundTrip(0xc001dc9530, 0xc0038a4200) vendor/k8s.io/client-go/transport/round_trippers.go:317 k8s.io/kubernetes/vendor/k8s.io/client-go/transport.(*userAgentRoundTripper).RoundTrip(0xc0016b67e0, 0xc0038a4100) vendor/k8s.io/client-go/transport/round_trippers.go:168 net/http.send(0xc0038a4100, {0x7fad100, 0xc0016b67e0}, {0x74d54e0?, 0x1?, 0x0?}) /usr/local/go/src/net/http/client.go:251 net/http.(*Client).send(0xc001dc9560, 0xc0038a4100, {0x7f9d4c43e108?, 0x100?, 0x0?}) /usr/local/go/src/net/http/client.go:175 net/http.(*Client).do(0xc001dc9560, 0xc0038a4100) /usr/local/go/src/net/http/client.go:715 net/http.(*Client).Do(...) /usr/local/go/src/net/http/client.go:581 k8s.io/kubernetes/vendor/k8s.io/client-go/rest.(*Request).request(0xc000ca1d00, {0x7fe0bc8, 0xc000136008}, 0x0?) vendor/k8s.io/client-go/rest/request.go:964 k8s.io/kubernetes/vendor/k8s.io/client-go/rest.(*Request).Do(0xc000ca1d00, {0x7fe0bc8, 0xc000136008}) vendor/k8s.io/client-go/rest/request.go:1005 k8s.io/kubernetes/vendor/k8s.io/client-go/kubernetes/typed/apps/v1.(*deployments).Get(0xc0018642e0, {0x7fe0bc8, 0xc000136008}, {0xc00557e570, 0x14}, {{{0x0, 0x0}, {0x0, 0x0}}, {0x0, ...}}) vendor/k8s.io/client-go/kubernetes/typed/apps/v1/deployment.go:86 k8s.io/kubernetes/test/utils.waitForDeploymentCompleteMaybeCheckRolling.func1() test/utils/deployment.go:84 k8s.io/kubernetes/vendor/k8s.io/apimachinery/pkg/util/wait.ConditionFunc.WithContext.func1({0x2742871, 0x0}) vendor/k8s.io/apimachinery/pkg/util/wait/wait.go:222 k8s.io/kubernetes/vendor/k8s.io/apimachinery/pkg/util/wait.runConditionWithCrashProtectionWithContext({0x7fe0bc8?, 0xc000136000?}, 0xc003fe7380?) vendor/k8s.io/apimachinery/pkg/util/wait/wait.go:235 k8s.io/kubernetes/vendor/k8s.io/apimachinery/pkg/util/wait.WaitForWithContext({0x7fe0bc8, 0xc000136000}, 0xc00557e5d0, 0x2fdb16a?) vendor/k8s.io/apimachinery/pkg/util/wait/wait.go:662 k8s.io/kubernetes/vendor/k8s.io/apimachinery/pkg/util/wait.poll({0x7fe0bc8, 0xc000136000}, 0x8?, 0x2fd9d05?, 0x40?) vendor/k8s.io/apimachinery/pkg/util/wait/wait.go:596 k8s.io/kubernetes/vendor/k8s.io/apimachinery/pkg/util/wait.PollImmediateWithContext({0x7fe0bc8, 0xc000136000}, 0x0?, 0xc00413bc58?, 0x262a967?) vendor/k8s.io/apimachinery/pkg/util/wait/wait.go:528 k8s.io/kubernetes/vendor/k8s.io/apimachinery/pkg/util/wait.PollImmediate(0x0?, 0x0?, 0x0?) vendor/k8s.io/apimachinery/pkg/util/wait/wait.go:514 k8s.io/kubernetes/test/utils.waitForDeploymentCompleteMaybeCheckRolling({0x801de88?, 0xc003fe7380}, 0xc000cc3900, 0x0, 0x78959b8, 0xa?, 0x1?) test/utils/deployment.go:82 k8s.io/kubernetes/test/utils.WaitForDeploymentComplete(...) test/utils/deployment.go:201 k8s.io/kubernetes/test/e2e/framework/deployment.WaitForDeploymentComplete(...) test/e2e/framework/deployment/wait.go:46 > k8s.io/kubernetes/test/e2e/network.glob..func20.6() test/e2e/network/loadbalancer.go:1444 k8s.io/kubernetes/vendor/github.com/onsi/ginkgo/v2/internal.extractBodyFunction.func3({0x2d591ce, 0xc00412a900}) vendor/github.com/onsi/ginkgo/v2/internal/node.go:449 k8s.io/kubernetes/vendor/github.com/onsi/ginkgo/v2/internal.(*Suite).runNode.func2() vendor/github.com/onsi/ginkgo/v2/internal/suite.go:750 k8s.io/kubernetes/vendor/github.com/onsi/ginkgo/v2/internal.(*Suite).runNode vendor/github.com/onsi/ginkgo/v2/internal/suite.go:738 ------------------------------ ------------------------------ Progress Report for Ginkgo Process #2 Automatically polling progress: [sig-network] LoadBalancers ESIPP [Slow] should work from pods (Spec Runtime: 12m38.667s) test/e2e/network/loadbalancer.go:1422 In [It] (Node Runtime: 10m40.036s) test/e2e/network/loadbalancer.go:1422 At [By Step] Creating pause pod deployment to make sure, pausePods are in desired state (Step Runtime: 2m51.389s) test/e2e/network/loadbalancer.go:1442 Spec Goroutine goroutine 2032 [select] k8s.io/kubernetes/vendor/golang.org/x/net/http2.(*ClientConn).RoundTrip(0xc000cec600, 0xc0038a4300) vendor/golang.org/x/net/http2/transport.go:1200 k8s.io/kubernetes/vendor/golang.org/x/net/http2.(*Transport).RoundTripOpt(0xc0040ec400, 0xc0038a4300, {0xe0?}) vendor/golang.org/x/net/http2/transport.go:519 k8s.io/kubernetes/vendor/golang.org/x/net/http2.(*Transport).RoundTrip(...) vendor/golang.org/x/net/http2/transport.go:480 k8s.io/kubernetes/vendor/golang.org/x/net/http2.noDialH2RoundTripper.RoundTrip({0xc0040f6000?}, 0xc0038a4300?) vendor/golang.org/x/net/http2/transport.go:3020 net/http.(*Transport).roundTrip(0xc0040f6000, 0xc0038a4300) /usr/local/go/src/net/http/transport.go:540 net/http.(*Transport).RoundTrip(0x6fe4b20?, 0xc0051f6810?) /usr/local/go/src/net/http/roundtrip.go:17 k8s.io/kubernetes/vendor/k8s.io/client-go/transport.(*bearerAuthRoundTripper).RoundTrip(0xc001dc9530, 0xc0038a4200) vendor/k8s.io/client-go/transport/round_trippers.go:317 k8s.io/kubernetes/vendor/k8s.io/client-go/transport.(*userAgentRoundTripper).RoundTrip(0xc0016b67e0, 0xc0038a4100) vendor/k8s.io/client-go/transport/round_trippers.go:168 net/http.send(0xc0038a4100, {0x7fad100, 0xc0016b67e0}, {0x74d54e0?, 0x1?, 0x0?}) /usr/local/go/src/net/http/client.go:251 net/http.(*Client).send(0xc001dc9560, 0xc0038a4100, {0x7f9d4c43e108?, 0x100?, 0x0?}) /usr/local/go/src/net/http/client.go:175 net/http.(*Client).do(0xc001dc9560, 0xc0038a4100) /usr/local/go/src/net/http/client.go:715 net/http.(*Client).Do(...) /usr/local/go/src/net/http/client.go:581 k8s.io/kubernetes/vendor/k8s.io/client-go/rest.(*Request).request(0xc000ca1d00, {0x7fe0bc8, 0xc000136008}, 0x0?) vendor/k8s.io/client-go/rest/request.go:964 k8s.io/kubernetes/vendor/k8s.io/client-go/rest.(*Request).Do(0xc000ca1d00, {0x7fe0bc8, 0xc000136008}) vendor/k8s.io/client-go/rest/request.go:1005 k8s.io/kubernetes/vendor/k8s.io/client-go/kubernetes/typed/apps/v1.(*deployments).Get(0xc0018642e0, {0x7fe0bc8, 0xc000136008}, {0xc00557e570, 0x14}, {{{0x0, 0x0}, {0x0, 0x0}}, {0x0, ...}}) vendor/k8s.io/client-go/kubernetes/typed/apps/v1/deployment.go:86 k8s.io/kubernetes/test/utils.waitForDeploymentCompleteMaybeCheckRolling.func1() test/utils/deployment.go:84 k8s.io/kubernetes/vendor/k8s.io/apimachinery/pkg/util/wait.ConditionFunc.WithContext.func1({0x2742871, 0x0}) vendor/k8s.io/apimachinery/pkg/util/wait/wait.go:222 k8s.io/kubernetes/vendor/k8s.io/apimachinery/pkg/util/wait.runConditionWithCrashProtectionWithContext({0x7fe0bc8?, 0xc000136000?}, 0xc003fe7380?) vendor/k8s.io/apimachinery/pkg/util/wait/wait.go:235 k8s.io/kubernetes/vendor/k8s.io/apimachinery/pkg/util/wait.WaitForWithContext({0x7fe0bc8, 0xc000136000}, 0xc00557e5d0, 0x2fdb16a?) vendor/k8s.io/apimachinery/pkg/util/wait/wait.go:662 k8s.io/kubernetes/vendor/k8s.io/apimachinery/pkg/util/wait.poll({0x7fe0bc8, 0xc000136000}, 0x8?, 0x2fd9d05?, 0x40?) vendor/k8s.io/apimachinery/pkg/util/wait/wait.go:596 k8s.io/kubernetes/vendor/k8s.io/apimachinery/pkg/util/wait.PollImmediateWithContext({0x7fe0bc8, 0xc000136000}, 0x0?, 0xc00413bc58?, 0x262a967?) vendor/k8s.io/apimachinery/pkg/util/wait/wait.go:528 k8s.io/kubernetes/vendor/k8s.io/apimachinery/pkg/util/wait.PollImmediate(0x0?, 0x0?, 0x0?) vendor/k8s.io/apimachinery/pkg/util/wait/wait.go:514 k8s.io/kubernetes/test/utils.waitForDeploymentCompleteMaybeCheckRolling({0x801de88?, 0xc003fe7380}, 0xc000cc3900, 0x0, 0x78959b8, 0xa?, 0x1?) test/utils/deployment.go:82 k8s.io/kubernetes/test/utils.WaitForDeploymentComplete(...) test/utils/deployment.go:201 k8s.io/kubernetes/test/e2e/framework/deployment.WaitForDeploymentComplete(...) test/e2e/framework/deployment/wait.go:46 > k8s.io/kubernetes/test/e2e/network.glob..func20.6() test/e2e/network/loadbalancer.go:1444 k8s.io/kubernetes/vendor/github.com/onsi/ginkgo/v2/internal.extractBodyFunction.func3({0x2d591ce, 0xc00412a900}) vendor/github.com/onsi/ginkgo/v2/internal/node.go:449 k8s.io/kubernetes/vendor/github.com/onsi/ginkgo/v2/internal.(*Suite).runNode.func2() vendor/github.com/onsi/ginkgo/v2/internal/suite.go:750 k8s.io/kubernetes/vendor/github.com/onsi/ginkgo/v2/internal.(*Suite).runNode vendor/github.com/onsi/ginkgo/v2/internal/suite.go:738 ------------------------------ Nov 26 05:26:41.561: INFO: Unexpected error: Failed to complete pause pod deployment: <*errors.errorString | 0xc00138c520>: { s: "error waiting for deployment \"pause-pod-deployment\" status to match expectation: Get \"https://35.227.179.144/apis/apps/v1/namespaces/esipp-9139/deployments/pause-pod-deployment\": stream error: stream ID 533; INTERNAL_ERROR; received from peer", } Nov 26 05:26:41.561: FAIL: Failed to complete pause pod deployment: error waiting for deployment "pause-pod-deployment" status to match expectation: Get "https://35.227.179.144/apis/apps/v1/namespaces/esipp-9139/deployments/pause-pod-deployment": stream error: stream ID 533; INTERNAL_ERROR; received from peer Full Stack Trace k8s.io/kubernetes/test/e2e/network.glob..func20.6() test/e2e/network/loadbalancer.go:1444 +0x3c6 ------------------------------ Progress Report for Ginkgo Process #2 Automatically polling progress: [sig-network] LoadBalancers ESIPP [Slow] should work from pods (Spec Runtime: 12m58.67s) test/e2e/network/loadbalancer.go:1422 In [It] (Node Runtime: 11m0.039s) test/e2e/network/loadbalancer.go:1422 At [By Step] Creating pause pod deployment to make sure, pausePods are in desired state (Step Runtime: 3m11.392s) test/e2e/network/loadbalancer.go:1442 Spec Goroutine goroutine 2032 [select] k8s.io/kubernetes/vendor/golang.org/x/net/http2.(*ClientConn).RoundTrip(0xc000cec600, 0xc0038a4800) vendor/golang.org/x/net/http2/transport.go:1200 k8s.io/kubernetes/vendor/golang.org/x/net/http2.(*Transport).RoundTripOpt(0xc0040ec400, 0xc0038a4800, {0xe0?}) vendor/golang.org/x/net/http2/transport.go:519 k8s.io/kubernetes/vendor/golang.org/x/net/http2.(*Transport).RoundTrip(...) vendor/golang.org/x/net/http2/transport.go:480 k8s.io/kubernetes/vendor/golang.org/x/net/http2.noDialH2RoundTripper.RoundTrip({0xc0040f6000?}, 0xc0038a4800?) vendor/golang.org/x/net/http2/transport.go:3020 net/http.(*Transport).roundTrip(0xc0040f6000, 0xc0038a4800) /usr/local/go/src/net/http/transport.go:540 net/http.(*Transport).RoundTrip(0x6fe4b20?, 0xc0051f6d80?) /usr/local/go/src/net/http/roundtrip.go:17 k8s.io/kubernetes/vendor/k8s.io/client-go/transport.(*bearerAuthRoundTripper).RoundTrip(0xc001dc9530, 0xc0038a4700) vendor/k8s.io/client-go/transport/round_trippers.go:317 k8s.io/kubernetes/vendor/k8s.io/client-go/transport.(*userAgentRoundTripper).RoundTrip(0xc0016b67e0, 0xc0038a4600) vendor/k8s.io/client-go/transport/round_trippers.go:168 net/http.send(0xc0038a4600, {0x7fad100, 0xc0016b67e0}, {0x74d54e0?, 0x1?, 0x0?}) /usr/local/go/src/net/http/client.go:251 net/http.(*Client).send(0xc001dc9560, 0xc0038a4600, {0x7f9d4c43e108?, 0x100?, 0x0?}) /usr/local/go/src/net/http/client.go:175 net/http.(*Client).do(0xc001dc9560, 0xc0038a4600) /usr/local/go/src/net/http/client.go:715 net/http.(*Client).Do(...) /usr/local/go/src/net/http/client.go:581 k8s.io/kubernetes/vendor/k8s.io/client-go/rest.(*Request).request(0xc0038a4400, {0x7fe0bc8, 0xc000136008}, 0x0?) vendor/k8s.io/client-go/rest/request.go:964 k8s.io/kubernetes/vendor/k8s.io/client-go/rest.(*Request).Do(0xc0038a4400, {0x7fe0bc8, 0xc000136008}) vendor/k8s.io/client-go/rest/request.go:1005 k8s.io/kubernetes/vendor/k8s.io/client-go/kubernetes/typed/core/v1.(*services).Get(0xc0048fe1e0, {0x7fe0bc8, 0xc000136008}, {0x75fa500, 0x13}, {{{0x0, 0x0}, {0x0, 0x0}}, {0x0, ...}}) vendor/k8s.io/client-go/kubernetes/typed/core/v1/service.go:79 k8s.io/kubernetes/test/e2e/framework/service.(*TestJig).UpdateService(0xc0045ff5e0, 0xc00413b880) test/e2e/framework/service/jig.go:523 k8s.io/kubernetes/test/e2e/framework/service.(*TestJig).ChangeServiceType(0x0?, {0x75c5095?, 0x0?}, 0x0?) test/e2e/framework/service/jig.go:173 > k8s.io/kubernetes/test/e2e/network.glob..func20.6.1() test/e2e/network/loadbalancer.go:1431 panic({0x70eb7e0, 0xc004950620}) /usr/local/go/src/runtime/panic.go:884 k8s.io/kubernetes/vendor/github.com/onsi/ginkgo/v2.Fail({0xc00051a3c0, 0x130}, {0xc00413bc20?, 0x75b521a?, 0xc00413bc40?}) vendor/github.com/onsi/ginkgo/v2/core_dsl.go:352 k8s.io/kubernetes/test/e2e/framework.Fail({0xc000ca8360, 0x11b}, {0xc00413bcb8?, 0xc00058c300?, 0xc00413bce0?}) test/e2e/framework/log.go:61 k8s.io/kubernetes/test/e2e/framework.ExpectNoErrorWithOffset(0x1, {0x7fa3ee0, 0xc00138c520}, {0xc00138c540?, 0x78959b8?, 0xa?}) test/e2e/framework/expect.go:76 k8s.io/kubernetes/test/e2e/framework.ExpectNoError(...) test/e2e/framework/expect.go:43 > k8s.io/kubernetes/test/e2e/network.glob..func20.6() test/e2e/network/loadbalancer.go:1444 k8s.io/kubernetes/vendor/github.com/onsi/ginkgo/v2/internal.extractBodyFunction.func3({0x2d591ce, 0xc00412a900}) vendor/github.com/onsi/ginkgo/v2/internal/node.go:449 k8s.io/kubernetes/vendor/github.com/onsi/ginkgo/v2/internal.(*Suite).runNode.func2() vendor/github.com/onsi/ginkgo/v2/internal/suite.go:750 k8s.io/kubernetes/vendor/github.com/onsi/ginkgo/v2/internal.(*Suite).runNode vendor/github.com/onsi/ginkgo/v2/internal/suite.go:738 ------------------------------ ------------------------------ Progress Report for Ginkgo Process #2 Automatically polling progress: [sig-network] LoadBalancers ESIPP [Slow] should work from pods (Spec Runtime: 13m18.673s) test/e2e/network/loadbalancer.go:1422 In [It] (Node Runtime: 11m20.042s) test/e2e/network/loadbalancer.go:1422 At [By Step] Creating pause pod deployment to make sure, pausePods are in desired state (Step Runtime: 3m31.395s) test/e2e/network/loadbalancer.go:1442 Spec Goroutine goroutine 2032 [select] k8s.io/kubernetes/vendor/golang.org/x/net/http2.(*ClientConn).RoundTrip(0xc000cec600, 0xc0038a4800) vendor/golang.org/x/net/http2/transport.go:1200 k8s.io/kubernetes/vendor/golang.org/x/net/http2.(*Transport).RoundTripOpt(0xc0040ec400, 0xc0038a4800, {0xe0?}) vendor/golang.org/x/net/http2/transport.go:519 k8s.io/kubernetes/vendor/golang.org/x/net/http2.(*Transport).RoundTrip(...) vendor/golang.org/x/net/http2/transport.go:480 k8s.io/kubernetes/vendor/golang.org/x/net/http2.noDialH2RoundTripper.RoundTrip({0xc0040f6000?}, 0xc0038a4800?) vendor/golang.org/x/net/http2/transport.go:3020 net/http.(*Transport).roundTrip(0xc0040f6000, 0xc0038a4800) /usr/local/go/src/net/http/transport.go:540 net/http.(*Transport).RoundTrip(0x6fe4b20?, 0xc0051f6d80?) /usr/local/go/src/net/http/roundtrip.go:17 k8s.io/kubernetes/vendor/k8s.io/client-go/transport.(*bearerAuthRoundTripper).RoundTrip(0xc001dc9530, 0xc0038a4700) vendor/k8s.io/client-go/transport/round_trippers.go:317 k8s.io/kubernetes/vendor/k8s.io/client-go/transport.(*userAgentRoundTripper).RoundTrip(0xc0016b67e0, 0xc0038a4600) vendor/k8s.io/client-go/transport/round_trippers.go:168 net/http.send(0xc0038a4600, {0x7fad100, 0xc0016b67e0}, {0x74d54e0?, 0x1?, 0x0?}) /usr/local/go/src/net/http/client.go:251 net/http.(*Client).send(0xc001dc9560, 0xc0038a4600, {0x7f9d4c43e108?, 0x100?, 0x0?}) /usr/local/go/src/net/http/client.go:175 net/http.(*Client).do(0xc001dc9560, 0xc0038a4600) /usr/local/go/src/net/http/client.go:715 net/http.(*Client).Do(...) /usr/local/go/src/net/http/client.go:581 k8s.io/kubernetes/vendor/k8s.io/client-go/rest.(*Request).request(0xc0038a4400, {0x7fe0bc8, 0xc000136008}, 0x0?) vendor/k8s.io/client-go/rest/request.go:964 k8s.io/kubernetes/vendor/k8s.io/client-go/rest.(*Request).Do(0xc0038a4400, {0x7fe0bc8, 0xc000136008}) vendor/k8s.io/client-go/rest/request.go:1005 k8s.io/kubernetes/vendor/k8s.io/client-go/kubernetes/typed/core/v1.(*services).Get(0xc0048fe1e0, {0x7fe0bc8, 0xc000136008}, {0x75fa500, 0x13}, {{{0x0, 0x0}, {0x0, 0x0}}, {0x0, ...}}) vendor/k8s.io/client-go/kubernetes/typed/core/v1/service.go:79 k8s.io/kubernetes/test/e2e/framework/service.(*TestJig).UpdateService(0xc0045ff5e0, 0xc00413b880) test/e2e/framework/service/jig.go:523 k8s.io/kubernetes/test/e2e/framework/service.(*TestJig).ChangeServiceType(0x0?, {0x75c5095?, 0x0?}, 0x0?) test/e2e/framework/service/jig.go:173 > k8s.io/kubernetes/test/e2e/network.glob..func20.6.1() test/e2e/network/loadbalancer.go:1431 panic({0x70eb7e0, 0xc004950620}) /usr/local/go/src/runtime/panic.go:884 k8s.io/kubernetes/vendor/github.com/onsi/ginkgo/v2.Fail({0xc00051a3c0, 0x130}, {0xc00413bc20?, 0x75b521a?, 0xc00413bc40?}) vendor/github.com/onsi/ginkgo/v2/core_dsl.go:352 k8s.io/kubernetes/test/e2e/framework.Fail({0xc000ca8360, 0x11b}, {0xc00413bcb8?, 0xc00058c300?, 0xc00413bce0?}) test/e2e/framework/log.go:61 k8s.io/kubernetes/test/e2e/framework.ExpectNoErrorWithOffset(0x1, {0x7fa3ee0, 0xc00138c520}, {0xc00138c540?, 0x78959b8?, 0xa?}) test/e2e/framework/expect.go:76 k8s.io/kubernetes/test/e2e/framework.ExpectNoError(...) test/e2e/framework/expect.go:43 > k8s.io/kubernetes/test/e2e/network.glob..func20.6() test/e2e/network/loadbalancer.go:1444 k8s.io/kubernetes/vendor/github.com/onsi/ginkgo/v2/internal.extractBodyFunction.func3({0x2d591ce, 0xc00412a900}) vendor/github.com/onsi/ginkgo/v2/internal/node.go:449 k8s.io/kubernetes/vendor/github.com/onsi/ginkgo/v2/internal.(*Suite).runNode.func2() vendor/github.com/onsi/ginkgo/v2/internal/suite.go:750 k8s.io/kubernetes/vendor/github.com/onsi/ginkgo/v2/internal.(*Suite).runNode vendor/github.com/onsi/ginkgo/v2/internal/suite.go:738 ------------------------------ ------------------------------ Progress Report for Ginkgo Process #2 Automatically polling progress: [sig-network] LoadBalancers ESIPP [Slow] should work from pods (Spec Runtime: 13m38.675s) test/e2e/network/loadbalancer.go:1422 In [It] (Node Runtime: 11m40.044s) test/e2e/network/loadbalancer.go:1422 At [By Step] Creating pause pod deployment to make sure, pausePods are in desired state (Step Runtime: 3m51.397s) test/e2e/network/loadbalancer.go:1442 Spec Goroutine goroutine 2032 [select, 2 minutes] k8s.io/kubernetes/vendor/golang.org/x/net/http2.(*ClientConn).RoundTrip(0xc000cec600, 0xc0038a4800) vendor/golang.org/x/net/http2/transport.go:1200 k8s.io/kubernetes/vendor/golang.org/x/net/http2.(*Transport).RoundTripOpt(0xc0040ec400, 0xc0038a4800, {0xe0?}) vendor/golang.org/x/net/http2/transport.go:519 k8s.io/kubernetes/vendor/golang.org/x/net/http2.(*Transport).RoundTrip(...) vendor/golang.org/x/net/http2/transport.go:480 k8s.io/kubernetes/vendor/golang.org/x/net/http2.noDialH2RoundTripper.RoundTrip({0xc0040f6000?}, 0xc0038a4800?) vendor/golang.org/x/net/http2/transport.go:3020 net/http.(*Transport).roundTrip(0xc0040f6000, 0xc0038a4800) /usr/local/go/src/net/http/transport.go:540 net/http.(*Transport).RoundTrip(0x6fe4b20?, 0xc0051f6d80?) /usr/local/go/src/net/http/roundtrip.go:17 k8s.io/kubernetes/vendor/k8s.io/client-go/transport.(*bearerAuthRoundTripper).RoundTrip(0xc001dc9530, 0xc0038a4700) vendor/k8s.io/client-go/transport/round_trippers.go:317 k8s.io/kubernetes/vendor/k8s.io/client-go/transport.(*userAgentRoundTripper).RoundTrip(0xc0016b67e0, 0xc0038a4600) vendor/k8s.io/client-go/transport/round_trippers.go:168 net/http.send(0xc0038a4600, {0x7fad100, 0xc0016b67e0}, {0x74d54e0?, 0x1?, 0x0?}) /usr/local/go/src/net/http/client.go:251 net/http.(*Client).send(0xc001dc9560, 0xc0038a4600, {0x7f9d4c43e108?, 0x100?, 0x0?}) /usr/local/go/src/net/http/client.go:175 net/http.(*Client).do(0xc001dc9560, 0xc0038a4600) /usr/local/go/src/net/http/client.go:715 net/http.(*Client).Do(...) /usr/local/go/src/net/http/client.go:581 k8s.io/kubernetes/vendor/k8s.io/client-go/rest.(*Request).request(0xc0038a4400, {0x7fe0bc8, 0xc000136008}, 0x0?) vendor/k8s.io/client-go/rest/request.go:964 k8s.io/kubernetes/vendor/k8s.io/client-go/rest.(*Request).Do(0xc0038a4400, {0x7fe0bc8, 0xc000136008}) vendor/k8s.io/client-go/rest/request.go:1005 k8s.io/kubernetes/vendor/k8s.io/client-go/kubernetes/typed/core/v1.(*services).Get(0xc0048fe1e0, {0x7fe0bc8, 0xc000136008}, {0x75fa500, 0x13}, {{{0x0, 0x0}, {0x0, 0x0}}, {0x0, ...}}) vendor/k8s.io/client-go/kubernetes/typed/core/v1/service.go:79 k8s.io/kubernetes/test/e2e/framework/service.(*TestJig).UpdateService(0xc0045ff5e0, 0xc00413b880) test/e2e/framework/service/jig.go:523 k8s.io/kubernetes/test/e2e/framework/service.(*TestJig).ChangeServiceType(0x0?, {0x75c5095?, 0x0?}, 0x0?) test/e2e/framework/service/jig.go:173 > k8s.io/kubernetes/test/e2e/network.glob..func20.6.1() test/e2e/network/loadbalancer.go:1431 panic({0x70eb7e0, 0xc004950620}) /usr/local/go/src/runtime/panic.go:884 k8s.io/kubernetes/vendor/github.com/onsi/ginkgo/v2.Fail({0xc00051a3c0, 0x130}, {0xc00413bc20?, 0x75b521a?, 0xc00413bc40?}) vendor/github.com/onsi/ginkgo/v2/core_dsl.go:352 k8s.io/kubernetes/test/e2e/framework.Fail({0xc000ca8360, 0x11b}, {0xc00413bcb8?, 0xc00058c300?, 0xc00413bce0?}) test/e2e/framework/log.go:61 k8s.io/kubernetes/test/e2e/framework.ExpectNoErrorWithOffset(0x1, {0x7fa3ee0, 0xc00138c520}, {0xc00138c540?, 0x78959b8?, 0xa?}) test/e2e/framework/expect.go:76 k8s.io/kubernetes/test/e2e/framework.ExpectNoError(...) test/e2e/framework/expect.go:43 > k8s.io/kubernetes/test/e2e/network.glob..func20.6() test/e2e/network/loadbalancer.go:1444 k8s.io/kubernetes/vendor/github.com/onsi/ginkgo/v2/internal.extractBodyFunction.func3({0x2d591ce, 0xc00412a900}) vendor/github.com/onsi/ginkgo/v2/internal/node.go:449 k8s.io/kubernetes/vendor/github.com/onsi/ginkgo/v2/internal.(*Suite).runNode.func2() vendor/github.com/onsi/ginkgo/v2/internal/suite.go:750 k8s.io/kubernetes/vendor/github.com/onsi/ginkgo/v2/internal.(*Suite).runNode vendor/github.com/onsi/ginkgo/v2/internal/suite.go:738 ------------------------------ Nov 26 05:27:40.515: INFO: Waiting up to 15m0s for service "external-local-pods" to have no LoadBalancer [AfterEach] [sig-network] LoadBalancers ESIPP [Slow] test/e2e/framework/node/init/init.go:32 Nov 26 05:27:50.774: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready [AfterEach] [sig-network] LoadBalancers ESIPP [Slow] test/e2e/network/loadbalancer.go:1260 Nov 26 05:27:50.821: INFO: Output of kubectl describe svc: Nov 26 05:27:50.822: INFO: Running '/workspace/github.com/containerd/containerd/kubernetes/platforms/linux/amd64/kubectl --server=https://35.227.179.144 --kubeconfig=/workspace/.kube/config --namespace=esipp-9139 describe svc --namespace=esipp-9139' Nov 26 05:27:51.166: INFO: stderr: "" Nov 26 05:27:51.166: INFO: stdout: "Name: external-local-pods\nNamespace: esipp-9139\nLabels: testid=external-local-pods-4f04647a-6eba-4ef7-bcdb-95258efd5a88\nAnnotations: <none>\nSelector: testid=external-local-pods-4f04647a-6eba-4ef7-bcdb-95258efd5a88\nType: ClusterIP\nIP Family Policy: SingleStack\nIP Families: IPv4\nIP: 10.0.170.167\nIPs: 10.0.170.167\nPort: <unset> 80/TCP\nTargetPort: 80/TCP\nEndpoints: 10.64.3.171:80\nSession Affinity: None\nEvents:\n Type Reason Age From Message\n ---- ------ ---- ---- -------\n Normal EnsuringLoadBalancer 10m service-controller Ensuring load balancer\n Normal EnsuringLoadBalancer 4m13s service-controller Ensuring load balancer\n Normal EnsuredLoadBalancer 4m12s service-controller Ensured load balancer\n" Nov 26 05:27:51.166: INFO: Name: external-local-pods Namespace: esipp-9139 Labels: testid=external-local-pods-4f04647a-6eba-4ef7-bcdb-95258efd5a88 Annotations: <none> Selector: testid=external-local-pods-4f04647a-6eba-4ef7-bcdb-95258efd5a88 Type: ClusterIP IP Family Policy: SingleStack IP Families: IPv4 IP: 10.0.170.167 IPs: 10.0.170.167 Port: <unset> 80/TCP TargetPort: 80/TCP Endpoints: 10.64.3.171:80 Session Affinity: None Events: Type Reason Age From Message ---- ------ ---- ---- ------- Normal EnsuringLoadBalancer 10m service-controller Ensuring load balancer Normal EnsuringLoadBalancer 4m13s service-controller Ensuring load balancer Normal EnsuredLoadBalancer 4m12s service-controller Ensured load balancer [DeferCleanup (Each)] [sig-network] LoadBalancers ESIPP [Slow] test/e2e/framework/metrics/init/init.go:33 [DeferCleanup (Each)] [sig-network] LoadBalancers ESIPP [Slow] dump namespaces | framework.go:196 STEP: dump namespace information after failure 11/26/22 05:27:51.166 STEP: Collecting events from namespace "esipp-9139". 11/26/22 05:27:51.166 STEP: Found 8 events. 11/26/22 05:27:51.208 Nov 26 05:27:51.208: INFO: At 0001-01-01 00:00:00 +0000 UTC - event for external-local-pods-4s55w: { } Scheduled: Successfully assigned esipp-9139/external-local-pods-4s55w to bootstrap-e2e-minion-group-vhdj Nov 26 05:27:51.208: INFO: At 2022-11-26 05:17:29 +0000 UTC - event for external-local-pods: {service-controller } EnsuringLoadBalancer: Ensuring load balancer Nov 26 05:27:51.208: INFO: At 2022-11-26 05:23:38 +0000 UTC - event for external-local-pods: {service-controller } EnsuringLoadBalancer: Ensuring load balancer Nov 26 05:27:51.208: INFO: At 2022-11-26 05:23:39 +0000 UTC - event for external-local-pods: {service-controller } EnsuredLoadBalancer: Ensured load balancer Nov 26 05:27:51.208: INFO: At 2022-11-26 05:23:41 +0000 UTC - event for external-local-pods: {replication-controller } SuccessfulCreate: Created pod: external-local-pods-4s55w Nov 26 05:27:51.208: INFO: At 2022-11-26 05:23:42 +0000 UTC - event for external-local-pods-4s55w: {kubelet bootstrap-e2e-minion-group-vhdj} Pulled: Container image "registry.k8s.io/e2e-test-images/agnhost:2.43" already present on machine Nov 26 05:27:51.208: INFO: At 2022-11-26 05:23:42 +0000 UTC - event for external-local-pods-4s55w: {kubelet bootstrap-e2e-minion-group-vhdj} Created: Created container netexec Nov 26 05:27:51.208: INFO: At 2022-11-26 05:23:42 +0000 UTC - event for external-local-pods-4s55w: {kubelet bootstrap-e2e-minion-group-vhdj} Started: Started container netexec Nov 26 05:27:51.250: INFO: POD NODE PHASE GRACE CONDITIONS Nov 26 05:27:51.250: INFO: external-local-pods-4s55w bootstrap-e2e-minion-group-vhdj Running [{Initialized True 0001-01-01 00:00:00 +0000 UTC 2022-11-26 05:23:41 +0000 UTC } {Ready True 0001-01-01 00:00:00 +0000 UTC 2022-11-26 05:23:43 +0000 UTC } {ContainersReady True 0001-01-01 00:00:00 +0000 UTC 2022-11-26 05:23:43 +0000 UTC } {PodScheduled True 0001-01-01 00:00:00 +0000 UTC 2022-11-26 05:23:41 +0000 UTC }] Nov 26 05:27:51.250: INFO: Nov 26 05:27:51.344: INFO: Logging node info for node bootstrap-e2e-master Nov 26 05:27:51.386: INFO: Node Info: &Node{ObjectMeta:{bootstrap-e2e-master 9a48eb11-c964-4bf8-82c0-1c32a9703981 10713 0 2022-11-26 05:05:18 +0000 UTC <nil> <nil> map[beta.kubernetes.io/arch:amd64 beta.kubernetes.io/instance-type:n1-standard-1 beta.kubernetes.io/os:linux cloud.google.com/metadata-proxy-ready:true failure-domain.beta.kubernetes.io/region:us-west1 failure-domain.beta.kubernetes.io/zone:us-west1-b kubernetes.io/arch:amd64 kubernetes.io/hostname:bootstrap-e2e-master kubernetes.io/os:linux node.kubernetes.io/instance-type:n1-standard-1 topology.kubernetes.io/region:us-west1 topology.kubernetes.io/zone:us-west1-b] map[node.alpha.kubernetes.io/ttl:0 volumes.kubernetes.io/controller-managed-attach-detach:true] [] [] [{kubelet Update v1 2022-11-26 05:05:18 +0000 UTC FieldsV1 {"f:metadata":{"f:annotations":{".":{},"f:volumes.kubernetes.io/controller-managed-attach-detach":{}},"f:labels":{".":{},"f:beta.kubernetes.io/arch":{},"f:beta.kubernetes.io/instance-type":{},"f:beta.kubernetes.io/os":{},"f:cloud.google.com/metadata-proxy-ready":{},"f:failure-domain.beta.kubernetes.io/region":{},"f:failure-domain.beta.kubernetes.io/zone":{},"f:kubernetes.io/arch":{},"f:kubernetes.io/hostname":{},"f:kubernetes.io/os":{},"f:node.kubernetes.io/instance-type":{},"f:topology.kubernetes.io/region":{},"f:topology.kubernetes.io/zone":{}}},"f:spec":{"f:providerID":{},"f:unschedulable":{}}} } {kube-controller-manager Update v1 2022-11-26 05:05:25 +0000 UTC FieldsV1 {"f:status":{"f:conditions":{"k:{\"type\":\"NetworkUnavailable\"}":{"f:lastHeartbeatTime":{},"f:lastTransitionTime":{},"f:message":{},"f:reason":{},"f:status":{}}}}} status} {kube-controller-manager Update v1 2022-11-26 05:05:39 +0000 UTC FieldsV1 {"f:metadata":{"f:annotations":{"f:node.alpha.kubernetes.io/ttl":{}}},"f:spec":{"f:podCIDR":{},"f:podCIDRs":{".":{},"v:\"10.64.1.0/24\"":{}},"f:taints":{}}} } {kubelet Update v1 2022-11-26 05:27:29 +0000 UTC FieldsV1 {"f:status":{"f:conditions":{"k:{\"type\":\"DiskPressure\"}":{"f:lastHeartbeatTime":{}},"k:{\"type\":\"MemoryPressure\"}":{"f:lastHeartbeatTime":{}},"k:{\"type\":\"PIDPressure\"}":{"f:lastHeartbeatTime":{}},"k:{\"type\":\"Ready\"}":{"f:lastHeartbeatTime":{},"f:lastTransitionTime":{},"f:message":{},"f:reason":{},"f:status":{}}},"f:images":{}}} status}]},Spec:NodeSpec{PodCIDR:10.64.1.0/24,DoNotUseExternalID:,ProviderID:gce://k8s-jkns-e2e-gce-ci-reboot/us-west1-b/bootstrap-e2e-master,Unschedulable:true,Taints:[]Taint{Taint{Key:node-role.kubernetes.io/master,Value:,Effect:NoSchedule,TimeAdded:<nil>,},Taint{Key:node.kubernetes.io/unschedulable,Value:,Effect:NoSchedule,TimeAdded:<nil>,},},ConfigSource:nil,PodCIDRs:[10.64.1.0/24],},Status:NodeStatus{Capacity:ResourceList{attachable-volumes-gce-pd: {{127 0} {<nil>} 127 DecimalSI},cpu: {{1 0} {<nil>} 1 DecimalSI},ephemeral-storage: {{16656896000 0} {<nil>} 16266500Ki BinarySI},hugepages-1Gi: {{0 0} {<nil>} 0 DecimalSI},hugepages-2Mi: {{0 0} {<nil>} 0 DecimalSI},memory: {{3858374656 0} {<nil>} BinarySI},pods: {{110 0} {<nil>} 110 DecimalSI},},Allocatable:ResourceList{attachable-volumes-gce-pd: {{127 0} {<nil>} 127 DecimalSI},cpu: {{1 0} {<nil>} 1 DecimalSI},ephemeral-storage: {{14991206376 0} {<nil>} 14991206376 DecimalSI},hugepages-1Gi: {{0 0} {<nil>} 0 DecimalSI},hugepages-2Mi: {{0 0} {<nil>} 0 DecimalSI},memory: {{3596230656 0} {<nil>} BinarySI},pods: {{110 0} {<nil>} 110 DecimalSI},},Phase:,Conditions:[]NodeCondition{NodeCondition{Type:NetworkUnavailable,Status:False,LastHeartbeatTime:2022-11-26 05:05:25 +0000 UTC,LastTransitionTime:2022-11-26 05:05:25 +0000 UTC,Reason:RouteCreated,Message:RouteController created a route,},NodeCondition{Type:MemoryPressure,Status:False,LastHeartbeatTime:2022-11-26 05:27:29 +0000 UTC,LastTransitionTime:2022-11-26 05:05:18 +0000 UTC,Reason:KubeletHasSufficientMemory,Message:kubelet has sufficient memory available,},NodeCondition{Type:DiskPressure,Status:False,LastHeartbeatTime:2022-11-26 05:27:29 +0000 UTC,LastTransitionTime:2022-11-26 05:05:18 +0000 UTC,Reason:KubeletHasNoDiskPressure,Message:kubelet has no disk pressure,},NodeCondition{Type:PIDPressure,Status:False,LastHeartbeatTime:2022-11-26 05:27:29 +0000 UTC,LastTransitionTime:2022-11-26 05:05:18 +0000 UTC,Reason:KubeletHasSufficientPID,Message:kubelet has sufficient PID available,},NodeCondition{Type:Ready,Status:True,LastHeartbeatTime:2022-11-26 05:27:29 +0000 UTC,LastTransitionTime:2022-11-26 05:05:38 +0000 UTC,Reason:KubeletReady,Message:kubelet is posting ready status. AppArmor enabled,},},Addresses:[]NodeAddress{NodeAddress{Type:InternalIP,Address:10.138.0.2,},NodeAddress{Type:ExternalIP,Address:35.227.179.144,},NodeAddress{Type:InternalDNS,Address:bootstrap-e2e-master.c.k8s-jkns-e2e-gce-ci-reboot.internal,},NodeAddress{Type:Hostname,Address:bootstrap-e2e-master.c.k8s-jkns-e2e-gce-ci-reboot.internal,},},DaemonEndpoints:NodeDaemonEndpoints{KubeletEndpoint:DaemonEndpoint{Port:10250,},},NodeInfo:NodeSystemInfo{MachineID:0fdb3cfe29f66637553465718381a2f8,SystemUUID:0fdb3cfe-29f6-6637-5534-65718381a2f8,BootID:fb804a29-7e61-434d-8044-6944159ce35f,KernelVersion:5.10.123+,OSImage:Container-Optimized OS from Google,ContainerRuntimeVersion:containerd://1.7.0-beta.0-149-gd06318622,KubeletVersion:v1.27.0-alpha.0.50+70617042976dc1,KubeProxyVersion:v1.27.0-alpha.0.50+70617042976dc1,OperatingSystem:linux,Architecture:amd64,},Images:[]ContainerImage{ContainerImage{Names:[registry.k8s.io/kube-apiserver-amd64:v1.27.0-alpha.0.50_70617042976dc1],SizeBytes:135160272,},ContainerImage{Names:[registry.k8s.io/kube-controller-manager-amd64:v1.27.0-alpha.0.50_70617042976dc1],SizeBytes:124990265,},ContainerImage{Names:[registry.k8s.io/etcd@sha256:dd75ec974b0a2a6f6bb47001ba09207976e625db898d1b16735528c009cb171c registry.k8s.io/etcd:3.5.6-0],SizeBytes:102542580,},ContainerImage{Names:[registry.k8s.io/kube-scheduler-amd64:v1.27.0-alpha.0.50_70617042976dc1],SizeBytes:57660216,},ContainerImage{Names:[gke.gcr.io/prometheus-to-sd@sha256:e739643c3939ba0b161425f45a1989eedfc4a3b166db9a7100863296b4c70510 gke.gcr.io/prometheus-to-sd:v0.11.1-gke.1],SizeBytes:48742566,},ContainerImage{Names:[gcr.io/k8s-ingress-image-push/ingress-gce-glbc-amd64@sha256:5db27383add6d9f4ebdf0286409ac31f7f5d273690204b341a4e37998917693b gcr.io/k8s-ingress-image-push/ingress-gce-glbc-amd64:v1.20.1],SizeBytes:36598135,},ContainerImage{Names:[registry.k8s.io/addon-manager/kube-addon-manager@sha256:49cc4e6e4a3745b427ce14b0141476ab339bb65c6bc05033019e046c8727dcb0 registry.k8s.io/addon-manager/kube-addon-manager:v9.1.6],SizeBytes:30464183,},ContainerImage{Names:[registry.k8s.io/kas-network-proxy/proxy-server@sha256:2c111f004bec24888d8cfa2a812a38fb8341350abac67dcd0ac64e709dfe389c registry.k8s.io/kas-network-proxy/proxy-server:v0.0.33],SizeBytes:22020129,},ContainerImage{Names:[registry.k8s.io/metadata-proxy@sha256:e914645f22e946bce5165737e1b244e0a296ad1f0f81a9531adc57af2780978a registry.k8s.io/metadata-proxy:v0.1.12],SizeBytes:5301657,},ContainerImage{Names:[registry.k8s.io/pause@sha256:9001185023633d17a2f98ff69b6ff2615b8ea02a825adffa40422f51dfdcde9d registry.k8s.io/pause:3.8],SizeBytes:311286,},},VolumesInUse:[],VolumesAttached:[]AttachedVolume{},Config:nil,},} Nov 26 05:27:51.386: INFO: Logging kubelet events for node bootstrap-e2e-master Nov 26 05:27:51.431: INFO: Logging pods the kubelet thinks is on node bootstrap-e2e-master Nov 26 05:27:51.486: INFO: metadata-proxy-v0.1-8ll2g started at 2022-11-26 05:05:51 +0000 UTC (0+2 container statuses recorded) Nov 26 05:27:51.486: INFO: Container metadata-proxy ready: true, restart count 0 Nov 26 05:27:51.486: INFO: Container prometheus-to-sd-exporter ready: true, restart count 0 Nov 26 05:27:51.486: INFO: etcd-server-events-bootstrap-e2e-master started at 2022-11-26 05:04:35 +0000 UTC (0+1 container statuses recorded) Nov 26 05:27:51.486: INFO: Container etcd-container ready: true, restart count 3 Nov 26 05:27:51.486: INFO: kube-controller-manager-bootstrap-e2e-master started at 2022-11-26 05:04:35 +0000 UTC (0+1 container statuses recorded) Nov 26 05:27:51.486: INFO: Container kube-controller-manager ready: false, restart count 6 Nov 26 05:27:51.486: INFO: kube-scheduler-bootstrap-e2e-master started at 2022-11-26 05:04:35 +0000 UTC (0+1 container statuses recorded) Nov 26 05:27:51.486: INFO: Container kube-scheduler ready: false, restart count 6 Nov 26 05:27:51.486: INFO: kube-addon-manager-bootstrap-e2e-master started at 2022-11-26 05:04:52 +0000 UTC (0+1 container statuses recorded) Nov 26 05:27:51.486: INFO: Container kube-addon-manager ready: true, restart count 1 Nov 26 05:27:51.486: INFO: etcd-server-bootstrap-e2e-master started at 2022-11-26 05:04:35 +0000 UTC (0+1 container statuses recorded) Nov 26 05:27:51.486: INFO: Container etcd-container ready: true, restart count 5 Nov 26 05:27:51.486: INFO: konnectivity-server-bootstrap-e2e-master started at 2022-11-26 05:04:35 +0000 UTC (0+1 container statuses recorded) Nov 26 05:27:51.486: INFO: Container konnectivity-server-container ready: true, restart count 2 Nov 26 05:27:51.486: INFO: kube-apiserver-bootstrap-e2e-master started at 2022-11-26 05:04:35 +0000 UTC (0+1 container statuses recorded) Nov 26 05:27:51.486: INFO: Container kube-apiserver ready: true, restart count 1 Nov 26 05:27:51.486: INFO: l7-lb-controller-bootstrap-e2e-master started at 2022-11-26 05:04:52 +0000 UTC (0+1 container statuses recorded) Nov 26 05:27:51.486: INFO: Container l7-lb-controller ready: false, restart count 7 Nov 26 05:27:51.712: INFO: Latency metrics for node bootstrap-e2e-master Nov 26 05:27:51.712: INFO: Logging node info for node bootstrap-e2e-minion-group-0975 Nov 26 05:27:51.755: INFO: Node Info: &Node{ObjectMeta:{bootstrap-e2e-minion-group-0975 6a0671d4-0e9d-4ac8-b359-65111262cfad 10851 0 2022-11-26 05:05:22 +0000 UTC <nil> <nil> map[beta.kubernetes.io/arch:amd64 beta.kubernetes.io/instance-type:n1-standard-2 beta.kubernetes.io/os:linux cloud.google.com/metadata-proxy-ready:true failure-domain.beta.kubernetes.io/region:us-west1 failure-domain.beta.kubernetes.io/zone:us-west1-b kubernetes.io/arch:amd64 kubernetes.io/hostname:bootstrap-e2e-minion-group-0975 kubernetes.io/os:linux node.kubernetes.io/instance-type:n1-standard-2 topology.hostpath.csi/node:bootstrap-e2e-minion-group-0975 topology.kubernetes.io/region:us-west1 topology.kubernetes.io/zone:us-west1-b] map[csi.volume.kubernetes.io/nodeid:{"csi-hostpath-multivolume-3400":"bootstrap-e2e-minion-group-0975","csi-hostpath-multivolume-5830":"bootstrap-e2e-minion-group-0975","csi-hostpath-multivolume-593":"bootstrap-e2e-minion-group-0975"} node.alpha.kubernetes.io/ttl:0 volumes.kubernetes.io/controller-managed-attach-detach:true] [] [] [{kubelet Update v1 2022-11-26 05:05:22 +0000 UTC FieldsV1 {"f:metadata":{"f:annotations":{".":{},"f:volumes.kubernetes.io/controller-managed-attach-detach":{}},"f:labels":{".":{},"f:beta.kubernetes.io/arch":{},"f:beta.kubernetes.io/instance-type":{},"f:beta.kubernetes.io/os":{},"f:cloud.google.com/metadata-proxy-ready":{},"f:failure-domain.beta.kubernetes.io/region":{},"f:failure-domain.beta.kubernetes.io/zone":{},"f:kubernetes.io/arch":{},"f:kubernetes.io/hostname":{},"f:kubernetes.io/os":{},"f:node.kubernetes.io/instance-type":{},"f:topology.kubernetes.io/region":{},"f:topology.kubernetes.io/zone":{}}},"f:spec":{"f:providerID":{}}} } {kube-controller-manager Update v1 2022-11-26 05:05:23 +0000 UTC FieldsV1 {"f:metadata":{"f:annotations":{"f:node.alpha.kubernetes.io/ttl":{}}},"f:spec":{"f:podCIDR":{},"f:podCIDRs":{".":{},"v:\"10.64.2.0/24\"":{}}}} } {kube-controller-manager Update v1 2022-11-26 05:13:31 +0000 UTC FieldsV1 {"f:status":{"f:conditions":{"k:{\"type\":\"NetworkUnavailable\"}":{"f:lastHeartbeatTime":{},"f:lastTransitionTime":{},"f:message":{},"f:reason":{},"f:status":{}}},"f:volumesAttached":{}}} status} {node-problem-detector Update v1 2022-11-26 05:25:30 +0000 UTC FieldsV1 {"f:status":{"f:conditions":{"k:{\"type\":\"CorruptDockerOverlay2\"}":{".":{},"f:lastHeartbeatTime":{},"f:lastTransitionTime":{},"f:message":{},"f:reason":{},"f:status":{},"f:type":{}},"k:{\"type\":\"FrequentContainerdRestart\"}":{".":{},"f:lastHeartbeatTime":{},"f:lastTransitionTime":{},"f:message":{},"f:reason":{},"f:status":{},"f:type":{}},"k:{\"type\":\"FrequentDockerRestart\"}":{".":{},"f:lastHeartbeatTime":{},"f:lastTransitionTime":{},"f:message":{},"f:reason":{},"f:status":{},"f:type":{}},"k:{\"type\":\"FrequentKubeletRestart\"}":{".":{},"f:lastHeartbeatTime":{},"f:lastTransitionTime":{},"f:message":{},"f:reason":{},"f:status":{},"f:type":{}},"k:{\"type\":\"FrequentUnregisterNetDevice\"}":{".":{},"f:lastHeartbeatTime":{},"f:lastTransitionTime":{},"f:message":{},"f:reason":{},"f:status":{},"f:type":{}},"k:{\"type\":\"KernelDeadlock\"}":{".":{},"f:lastHeartbeatTime":{},"f:lastTransitionTime":{},"f:message":{},"f:reason":{},"f:status":{},"f:type":{}},"k:{\"type\":\"ReadonlyFilesystem\"}":{".":{},"f:lastHeartbeatTime":{},"f:lastTransitionTime":{},"f:message":{},"f:reason":{},"f:status":{},"f:type":{}}}}} status} {kubelet Update v1 2022-11-26 05:27:46 +0000 UTC FieldsV1 {"f:metadata":{"f:annotations":{"f:csi.volume.kubernetes.io/nodeid":{}},"f:labels":{"f:topology.hostpath.csi/node":{}}},"f:status":{"f:conditions":{"k:{\"type\":\"DiskPressure\"}":{"f:lastHeartbeatTime":{}},"k:{\"type\":\"MemoryPressure\"}":{"f:lastHeartbeatTime":{}},"k:{\"type\":\"PIDPressure\"}":{"f:lastHeartbeatTime":{}},"k:{\"type\":\"Ready\"}":{"f:lastHeartbeatTime":{},"f:lastTransitionTime":{},"f:message":{},"f:reason":{},"f:status":{}}},"f:images":{}}} status}]},Spec:NodeSpec{PodCIDR:10.64.2.0/24,DoNotUseExternalID:,ProviderID:gce://k8s-jkns-e2e-gce-ci-reboot/us-west1-b/bootstrap-e2e-minion-group-0975,Unschedulable:false,Taints:[]Taint{},ConfigSource:nil,PodCIDRs:[10.64.2.0/24],},Status:NodeStatus{Capacity:ResourceList{attachable-volumes-gce-pd: {{127 0} {<nil>} 127 DecimalSI},cpu: {{2 0} {<nil>} 2 DecimalSI},ephemeral-storage: {{101203873792 0} {<nil>} 98831908Ki BinarySI},hugepages-1Gi: {{0 0} {<nil>} 0 DecimalSI},hugepages-2Mi: {{0 0} {<nil>} 0 DecimalSI},memory: {{7815430144 0} {<nil>} BinarySI},pods: {{110 0} {<nil>} 110 DecimalSI},},Allocatable:ResourceList{attachable-volumes-gce-pd: {{127 0} {<nil>} 127 DecimalSI},cpu: {{2 0} {<nil>} 2 DecimalSI},ephemeral-storage: {{91083486262 0} {<nil>} 91083486262 DecimalSI},hugepages-1Gi: {{0 0} {<nil>} 0 DecimalSI},hugepages-2Mi: {{0 0} {<nil>} 0 DecimalSI},memory: {{7553286144 0} {<nil>} BinarySI},pods: {{110 0} {<nil>} 110 DecimalSI},},Phase:,Conditions:[]NodeCondition{NodeCondition{Type:ReadonlyFilesystem,Status:False,LastHeartbeatTime:2022-11-26 05:25:30 +0000 UTC,LastTransitionTime:2022-11-26 05:05:27 +0000 UTC,Reason:FilesystemIsNotReadOnly,Message:Filesystem is not read-only,},NodeCondition{Type:CorruptDockerOverlay2,Status:False,LastHeartbeatTime:2022-11-26 05:25:30 +0000 UTC,LastTransitionTime:2022-11-26 05:05:27 +0000 UTC,Reason:NoCorruptDockerOverlay2,Message:docker overlay2 is functioning properly,},NodeCondition{Type:FrequentUnregisterNetDevice,Status:False,LastHeartbeatTime:2022-11-26 05:25:30 +0000 UTC,LastTransitionTime:2022-11-26 05:05:27 +0000 UTC,Reason:NoFrequentUnregisterNetDevice,Message:node is functioning properly,},NodeCondition{Type:FrequentKubeletRestart,Status:False,LastHeartbeatTime:2022-11-26 05:25:30 +0000 UTC,LastTransitionTime:2022-11-26 05:05:27 +0000 UTC,Reason:NoFrequentKubeletRestart,Message:kubelet is functioning properly,},NodeCondition{Type:FrequentDockerRestart,Status:False,LastHeartbeatTime:2022-11-26 05:25:30 +0000 UTC,LastTransitionTime:2022-11-26 05:05:27 +0000 UTC,Reason:NoFrequentDockerRestart,Message:docker is functioning properly,},NodeCondition{Type:FrequentContainerdRestart,Status:False,LastHeartbeatTime:2022-11-26 05:25:30 +0000 UTC,LastTransitionTime:2022-11-26 05:05:27 +0000 UTC,Reason:NoFrequentContainerdRestart,Message:containerd is functioning properly,},NodeCondition{Type:KernelDeadlock,Status:False,LastHeartbeatTime:2022-11-26 05:25:30 +0000 UTC,LastTransitionTime:2022-11-26 05:05:27 +0000 UTC,Reason:KernelHasNoDeadlock,Message:kernel has no deadlock,},NodeCondition{Type:NetworkUnavailable,Status:False,LastHeartbeatTime:2022-11-26 05:05:39 +0000 UTC,LastTransitionTime:2022-11-26 05:05:39 +0000 UTC,Reason:RouteCreated,Message:RouteController created a route,},NodeCondition{Type:MemoryPressure,Status:False,LastHeartbeatTime:2022-11-26 05:24:22 +0000 UTC,LastTransitionTime:2022-11-26 05:05:22 +0000 UTC,Reason:KubeletHasSufficientMemory,Message:kubelet has sufficient memory available,},NodeCondition{Type:DiskPressure,Status:False,LastHeartbeatTime:2022-11-26 05:24:22 +0000 UTC,LastTransitionTime:2022-11-26 05:05:22 +0000 UTC,Reason:KubeletHasNoDiskPressure,Message:kubelet has no disk pressure,},NodeCondition{Type:PIDPressure,Status:False,LastHeartbeatTime:2022-11-26 05:24:22 +0000 UTC,LastTransitionTime:2022-11-26 05:05:22 +0000 UTC,Reason:KubeletHasSufficientPID,Message:kubelet has sufficient PID available,},NodeCondition{Type:Ready,Status:True,LastHeartbeatTime:2022-11-26 05:24:22 +0000 UTC,LastTransitionTime:2022-11-26 05:05:23 +0000 UTC,Reason:KubeletReady,Message:kubelet is posting ready status. AppArmor enabled,},},Addresses:[]NodeAddress{NodeAddress{Type:InternalIP,Address:10.138.0.5,},NodeAddress{Type:ExternalIP,Address:35.247.104.236,},NodeAddress{Type:InternalDNS,Address:bootstrap-e2e-minion-group-0975.c.k8s-jkns-e2e-gce-ci-reboot.internal,},NodeAddress{Type:Hostname,Address:bootstrap-e2e-minion-group-0975.c.k8s-jkns-e2e-gce-ci-reboot.internal,},},DaemonEndpoints:NodeDaemonEndpoints{KubeletEndpoint:DaemonEndpoint{Port:10250,},},NodeInfo:NodeSystemInfo{MachineID:b38385f7976a474936da82f5f565760a,SystemUUID:b38385f7-976a-4749-36da-82f5f565760a,BootID:0aa74273-9b8d-48ac-bad0-0a1805f010ea,KernelVersion:5.10.123+,OSImage:Container-Optimized OS from Google,ContainerRuntimeVersion:containerd://1.7.0-beta.0-149-gd06318622,KubeletVersion:v1.27.0-alpha.0.50+70617042976dc1,KubeProxyVersion:v1.27.0-alpha.0.50+70617042976dc1,OperatingSystem:linux,Architecture:amd64,},Images:[]ContainerImage{ContainerImage{Names:[registry.k8s.io/e2e-test-images/jessie-dnsutils@sha256:24aaf2626d6b27864c29de2097e8bbb840b3a414271bf7c8995e431e47d8408e registry.k8s.io/e2e-test-images/jessie-dnsutils:1.7],SizeBytes:112030336,},ContainerImage{Names:[registry.k8s.io/e2e-test-images/volume/nfs@sha256:3bda73f2428522b0e342af80a0b9679e8594c2126f2b3cca39ed787589741b9e registry.k8s.io/e2e-test-images/volume/nfs:1.3],SizeBytes:95836203,},ContainerImage{Names:[registry.k8s.io/sig-storage/nfs-provisioner@sha256:e943bb77c7df05ebdc8c7888b2db289b13bf9f012d6a3a5a74f14d4d5743d439 registry.k8s.io/sig-storage/nfs-provisioner:v3.0.1],SizeBytes:90632047,},ContainerImage{Names:[registry.k8s.io/kube-proxy-amd64:v1.27.0-alpha.0.50_70617042976dc1],SizeBytes:67201736,},ContainerImage{Names:[registry.k8s.io/e2e-test-images/agnhost@sha256:16bbf38c463a4223d8cfe4da12bc61010b082a79b4bb003e2d3ba3ece5dd5f9e registry.k8s.io/e2e-test-images/agnhost:2.43],SizeBytes:51706353,},ContainerImage{Names:[gke.gcr.io/prometheus-to-sd@sha256:e739643c3939ba0b161425f45a1989eedfc4a3b166db9a7100863296b4c70510 gke.gcr.io/prometheus-to-sd:v0.11.1-gke.1],SizeBytes:48742566,},ContainerImage{Names:[registry.k8s.io/e2e-test-images/httpd@sha256:148b022f5c5da426fc2f3c14b5c0867e58ef05961510c84749ac1fddcb0fef22 registry.k8s.io/e2e-test-images/httpd:2.4.38-4],SizeBytes:40764257,},ContainerImage{Names:[registry.k8s.io/metrics-server/metrics-server@sha256:6385aec64bb97040a5e692947107b81e178555c7a5b71caa90d733e4130efc10 registry.k8s.io/metrics-server/metrics-server:v0.5.2],SizeBytes:26023008,},ContainerImage{Names:[registry.k8s.io/sig-storage/csi-provisioner@sha256:ee3b525d5b89db99da3b8eb521d9cd90cb6e9ef0fbb651e98bb37be78d36b5b8 registry.k8s.io/sig-storage/csi-provisioner:v3.3.0],SizeBytes:25491225,},ContainerImage{Names:[registry.k8s.io/sig-storage/csi-resizer@sha256:425d8f1b769398127767b06ed97ce62578a3179bcb99809ce93a1649e025ffe7 registry.k8s.io/sig-storage/csi-resizer:v1.6.0],SizeBytes:24148884,},ContainerImage{Names:[registry.k8s.io/sig-storage/csi-snapshotter@sha256:291334908ddf71a4661fd7f6d9d97274de8a5378a2b6fdfeb2ce73414a34f82f registry.k8s.io/sig-storage/csi-snapshotter:v6.1.0],SizeBytes:23881995,},ContainerImage{Names:[registry.k8s.io/sig-storage/csi-attacher@sha256:9a685020911e2725ad019dbce6e4a5ab93d51e3d4557f115e64343345e05781b registry.k8s.io/sig-storage/csi-attacher:v4.0.0],SizeBytes:23847201,},ContainerImage{Names:[registry.k8s.io/sig-storage/hostpathplugin@sha256:92257881c1d6493cf18299a24af42330f891166560047902b8d431fb66b01af5 registry.k8s.io/sig-storage/hostpathplugin:v1.9.0],SizeBytes:18758628,},ContainerImage{Names:[registry.k8s.io/autoscaling/addon-resizer@sha256:43f129b81d28f0fdd54de6d8e7eacd5728030782e03db16087fc241ad747d3d6 registry.k8s.io/autoscaling/addon-resizer:1.8.14],SizeBytes:10153852,},ContainerImage{Names:[registry.k8s.io/sig-storage/csi-node-driver-registrar@sha256:0103eee7c35e3e0b5cd8cdca9850dc71c793cdeb6669d8be7a89440da2d06ae4 registry.k8s.io/sig-storage/csi-node-driver-registrar:v2.5.1],SizeBytes:9133109,},ContainerImage{Names:[registry.k8s.io/sig-storage/livenessprobe@sha256:933940f13b3ea0abc62e656c1aa5c5b47c04b15d71250413a6b821bd0c58b94e registry.k8s.io/sig-storage/livenessprobe:v2.7.0],SizeBytes:8688564,},ContainerImage{Names:[registry.k8s.io/kas-network-proxy/proxy-agent@sha256:48f2a4ec3e10553a81b8dd1c6fa5fe4bcc9617f78e71c1ca89c6921335e2d7da registry.k8s.io/kas-network-proxy/proxy-agent:v0.0.33],SizeBytes:8512162,},ContainerImage{Names:[registry.k8s.io/metadata-proxy@sha256:e914645f22e946bce5165737e1b244e0a296ad1f0f81a9531adc57af2780978a registry.k8s.io/metadata-proxy:v0.1.12],SizeBytes:5301657,},ContainerImage{Names:[registry.k8s.io/e2e-test-images/busybox@sha256:c318242786b139d18676b1c09a0ad7f15fc17f8f16a5b2e625cd0dc8c9703daf registry.k8s.io/e2e-test-images/busybox:1.29-2],SizeBytes:732424,},ContainerImage{Names:[registry.k8s.io/e2e-test-images/busybox@sha256:2e0f836850e09b8b7cc937681d6194537a09fbd5f6b9e08f4d646a85128e8937 registry.k8s.io/e2e-test-images/busybox:1.29-4],SizeBytes:731990,},ContainerImage{Names:[registry.k8s.io/pause@sha256:9001185023633d17a2f98ff69b6ff2615b8ea02a825adffa40422f51dfdcde9d registry.k8s.io/pause:3.8],SizeBytes:311286,},},VolumesInUse:[],VolumesAttached:[]AttachedVolume{AttachedVolume{Name:kubernetes.io/csi/csi-hostpath-multivolume-702^02fad0b0-6d49-11ed-9d4c-26ec58c28662,DevicePath:,},},Config:nil,},} Nov 26 05:27:51.755: INFO: Logging kubelet events for node bootstrap-e2e-minion-group-0975 Nov 26 05:27:51.799: INFO: Logging pods the kubelet thinks is on node bootstrap-e2e-minion-group-0975 Nov 26 05:27:51.898: INFO: external-local-update-5xshb started at 2022-11-26 05:11:23 +0000 UTC (0+1 container statuses recorded) Nov 26 05:27:51.898: INFO: Container netexec ready: false, restart count 7 Nov 26 05:27:51.898: INFO: host-test-container-pod started at 2022-11-26 05:12:10 +0000 UTC (0+1 container statuses recorded) Nov 26 05:27:51.898: INFO: Container agnhost-container ready: true, restart count 4 Nov 26 05:27:51.898: INFO: forbid-27824003-x9rls started at 2022-11-26 05:23:10 +0000 UTC (0+1 container statuses recorded) Nov 26 05:27:51.898: INFO: Container c ready: true, restart count 0 Nov 26 05:27:51.898: INFO: metadata-proxy-v0.1-mznrc started at 2022-11-26 05:05:24 +0000 UTC (0+2 container statuses recorded) Nov 26 05:27:51.898: INFO: Container metadata-proxy ready: true, restart count 0 Nov 26 05:27:51.898: INFO: Container prometheus-to-sd-exporter ready: true, restart count 0 Nov 26 05:27:51.898: INFO: konnectivity-agent-wfc29 started at 2022-11-26 05:05:39 +0000 UTC (0+1 container statuses recorded) Nov 26 05:27:51.898: INFO: Container konnectivity-agent ready: true, restart count 7 Nov 26 05:27:51.898: INFO: csi-hostpathplugin-0 started at 2022-11-26 05:09:53 +0000 UTC (0+7 container statuses recorded) Nov 26 05:27:51.898: INFO: Container csi-attacher ready: true, restart count 3 Nov 26 05:27:51.898: INFO: Container csi-provisioner ready: true, restart count 3 Nov 26 05:27:51.898: INFO: Container csi-resizer ready: true, restart count 3 Nov 26 05:27:51.898: INFO: Container csi-snapshotter ready: true, restart count 3 Nov 26 05:27:51.898: INFO: Container hostpath ready: true, restart count 3 Nov 26 05:27:51.898: INFO: Container liveness-probe ready: true, restart count 3 Nov 26 05:27:51.898: INFO: Container node-driver-registrar ready: true, restart count 3 Nov 26 05:27:51.898: INFO: csi-mockplugin-0 started at 2022-11-26 05:23:18 +0000 UTC (0+4 container statuses recorded) Nov 26 05:27:51.898: INFO: Container busybox ready: false, restart count 3 Nov 26 05:27:51.898: INFO: Container csi-provisioner ready: true, restart count 4 Nov 26 05:27:51.898: INFO: Container driver-registrar ready: true, restart count 4 Nov 26 05:27:51.898: INFO: Container mock ready: true, restart count 4 Nov 26 05:27:51.898: INFO: csi-hostpathplugin-0 started at 2022-11-26 05:12:10 +0000 UTC (0+7 container statuses recorded) Nov 26 05:27:51.898: INFO: Container csi-attacher ready: false, restart count 5 Nov 26 05:27:51.898: INFO: Container csi-provisioner ready: false, restart count 5 Nov 26 05:27:51.898: INFO: Container csi-resizer ready: false, restart count 5 Nov 26 05:27:51.898: INFO: Container csi-snapshotter ready: false, restart count 5 Nov 26 05:27:51.898: INFO: Container hostpath ready: false, restart count 5 Nov 26 05:27:51.898: INFO: Container liveness-probe ready: false, restart count 5 Nov 26 05:27:51.898: INFO: Container node-driver-registrar ready: false, restart count 5 Nov 26 05:27:51.898: INFO: csi-hostpathplugin-0 started at 2022-11-26 05:23:17 +0000 UTC (0+7 container statuses recorded) Nov 26 05:27:51.898: INFO: Container csi-attacher ready: true, restart count 2 Nov 26 05:27:51.898: INFO: Container csi-provisioner ready: true, restart count 2 Nov 26 05:27:51.898: INFO: Container csi-resizer ready: true, restart count 2 Nov 26 05:27:51.898: INFO: Container csi-snapshotter ready: true, restart count 2 Nov 26 05:27:51.898: INFO: Container hostpath ready: true, restart count 2 Nov 26 05:27:51.898: INFO: Container liveness-probe ready: true, restart count 2 Nov 26 05:27:51.898: INFO: Container node-driver-registrar ready: true, restart count 2 Nov 26 05:27:51.898: INFO: pod-secrets-ae6e2110-abaa-4433-ab33-c417114f813f started at 2022-11-26 05:10:35 +0000 UTC (0+1 container statuses recorded) Nov 26 05:27:51.898: INFO: Container creates-volume-test ready: false, restart count 0 Nov 26 05:27:51.898: INFO: csi-hostpathplugin-0 started at 2022-11-26 05:13:04 +0000 UTC (0+7 container statuses recorded) Nov 26 05:27:51.898: INFO: Container csi-attacher ready: true, restart count 3 Nov 26 05:27:51.898: INFO: Container csi-provisioner ready: true, restart count 3 Nov 26 05:27:51.898: INFO: Container csi-resizer ready: true, restart count 3 Nov 26 05:27:51.898: INFO: Container csi-snapshotter ready: true, restart count 3 Nov 26 05:27:51.898: INFO: Container hostpath ready: true, restart count 3 Nov 26 05:27:51.898: INFO: Container liveness-probe ready: true, restart count 3 Nov 26 05:27:51.898: INFO: Container node-driver-registrar ready: true, restart count 3 Nov 26 05:27:51.898: INFO: netserver-0 started at 2022-11-26 05:11:06 +0000 UTC (0+1 container statuses recorded) Nov 26 05:27:51.898: INFO: Container webserver ready: false, restart count 6 Nov 26 05:27:51.898: INFO: kube-proxy-bootstrap-e2e-minion-group-0975 started at 2022-11-26 05:05:23 +0000 UTC (0+1 container statuses recorded) Nov 26 05:27:51.898: INFO: Container kube-proxy ready: false, restart count 7 Nov 26 05:27:51.898: INFO: netserver-0 started at 2022-11-26 05:11:29 +0000 UTC (0+1 container statuses recorded) Nov 26 05:27:51.898: INFO: Container webserver ready: true, restart count 6 Nov 26 05:27:51.898: INFO: hostexec-bootstrap-e2e-minion-group-0975-xk996 started at 2022-11-26 05:23:32 +0000 UTC (0+1 container statuses recorded) Nov 26 05:27:51.898: INFO: Container agnhost-container ready: true, restart count 0 Nov 26 05:27:51.898: INFO: affinity-lb-esipp-2xp2g started at 2022-11-26 05:23:40 +0000 UTC (0+1 container statuses recorded) Nov 26 05:27:51.898: INFO: Container affinity-lb-esipp ready: true, restart count 1 Nov 26 05:27:51.898: INFO: metrics-server-v0.5.2-867b8754b9-27xcg started at 2022-11-26 05:05:51 +0000 UTC (0+2 container statuses recorded) Nov 26 05:27:51.898: INFO: Container metrics-server ready: false, restart count 7 Nov 26 05:27:51.898: INFO: Container metrics-server-nanny ready: false, restart count 8 Nov 26 05:27:51.898: INFO: test-hostpath-type-2p74k started at 2022-11-26 05:17:26 +0000 UTC (0+1 container statuses recorded) Nov 26 05:27:51.898: INFO: Container host-path-testing ready: false, restart count 0 Nov 26 05:27:51.898: INFO: nfs-server started at 2022-11-26 05:23:15 +0000 UTC (0+1 container statuses recorded) Nov 26 05:27:51.898: INFO: Container nfs-server ready: true, restart count 1 Nov 26 05:27:51.898: INFO: net-tiers-svc-sj7rf started at 2022-11-26 05:17:11 +0000 UTC (0+1 container statuses recorded) Nov 26 05:27:51.898: INFO: Container netexec ready: true, restart count 1 Nov 26 05:27:51.898: INFO: reallocate-nodeport-test-njbll started at 2022-11-26 05:23:12 +0000 UTC (0+1 container statuses recorded) Nov 26 05:27:51.898: INFO: Container netexec ready: true, restart count 2 Nov 26 05:27:51.898: INFO: hostexec-bootstrap-e2e-minion-group-0975-vp665 started at 2022-11-26 05:17:32 +0000 UTC (0+1 container statuses recorded) Nov 26 05:27:51.898: INFO: Container agnhost-container ready: true, restart count 1 Nov 26 05:27:51.898: INFO: lb-internal-r5l4p started at 2022-11-26 05:16:06 +0000 UTC (0+1 container statuses recorded) Nov 26 05:27:51.898: INFO: Container netexec ready: true, restart count 0 Nov 26 05:27:51.898: INFO: csi-mockplugin-0 started at 2022-11-26 05:17:37 +0000 UTC (0+4 container statuses recorded) Nov 26 05:27:51.898: INFO: Container busybox ready: false, restart count 4 Nov 26 05:27:51.898: INFO: Container csi-provisioner ready: false, restart count 3 Nov 26 05:27:51.898: INFO: Container driver-registrar ready: false, restart count 4 Nov 26 05:27:51.898: INFO: Container mock ready: false, restart count 4 Nov 26 05:27:51.898: INFO: netserver-0 started at 2022-11-26 05:09:51 +0000 UTC (0+1 container statuses recorded) Nov 26 05:27:51.898: INFO: Container webserver ready: true, restart count 6 Nov 26 05:27:51.898: INFO: hostexec-bootstrap-e2e-minion-group-0975-4z5l4 started at 2022-11-26 05:23:27 +0000 UTC (0+1 container statuses recorded) Nov 26 05:27:51.898: INFO: Container agnhost-container ready: true, restart count 1 Nov 26 05:27:52.206: INFO: Latency metrics for node bootstrap-e2e-minion-group-0975 Nov 26 05:27:52.206: INFO: Logging node info for node bootstrap-e2e-minion-group-rwsn Nov 26 05:27:52.248: INFO: Node Info: &Node{ObjectMeta:{bootstrap-e2e-minion-group-rwsn adedf225-b462-4305-a60c-9042d31ca1d2 10874 0 2022-11-26 05:05:15 +0000 UTC <nil> <nil> map[beta.kubernetes.io/arch:amd64 beta.kubernetes.io/instance-type:n1-standard-2 beta.kubernetes.io/os:linux cloud.google.com/metadata-proxy-ready:true failure-domain.beta.kubernetes.io/region:us-west1 failure-domain.beta.kubernetes.io/zone:us-west1-b kubernetes.io/arch:amd64 kubernetes.io/hostname:bootstrap-e2e-minion-group-rwsn kubernetes.io/os:linux node.kubernetes.io/instance-type:n1-standard-2 topology.hostpath.csi/node:bootstrap-e2e-minion-group-rwsn topology.kubernetes.io/region:us-west1 topology.kubernetes.io/zone:us-west1-b] map[node.alpha.kubernetes.io/ttl:0 volumes.kubernetes.io/controller-managed-attach-detach:true] [] [] [{kubelet Update v1 2022-11-26 05:05:15 +0000 UTC FieldsV1 {"f:metadata":{"f:annotations":{".":{},"f:volumes.kubernetes.io/controller-managed-attach-detach":{}},"f:labels":{".":{},"f:beta.kubernetes.io/arch":{},"f:beta.kubernetes.io/instance-type":{},"f:beta.kubernetes.io/os":{},"f:cloud.google.com/metadata-proxy-ready":{},"f:failure-domain.beta.kubernetes.io/region":{},"f:failure-domain.beta.kubernetes.io/zone":{},"f:kubernetes.io/arch":{},"f:kubernetes.io/hostname":{},"f:kubernetes.io/os":{},"f:node.kubernetes.io/instance-type":{},"f:topology.kubernetes.io/region":{},"f:topology.kubernetes.io/zone":{}}},"f:spec":{"f:providerID":{}}} } {kube-controller-manager Update v1 2022-11-26 05:05:17 +0000 UTC FieldsV1 {"f:metadata":{"f:annotations":{"f:node.alpha.kubernetes.io/ttl":{}}},"f:spec":{"f:podCIDR":{},"f:podCIDRs":{".":{},"v:\"10.64.0.0/24\"":{}}}} } {kube-controller-manager Update v1 2022-11-26 05:23:27 +0000 UTC FieldsV1 {"f:status":{"f:conditions":{"k:{\"type\":\"NetworkUnavailable\"}":{"f:lastHeartbeatTime":{},"f:lastTransitionTime":{},"f:message":{},"f:reason":{},"f:status":{}}},"f:volumesAttached":{}}} status} {node-problem-detector Update v1 2022-11-26 05:25:21 +0000 UTC FieldsV1 {"f:status":{"f:conditions":{"k:{\"type\":\"CorruptDockerOverlay2\"}":{".":{},"f:lastHeartbeatTime":{},"f:lastTransitionTime":{},"f:message":{},"f:reason":{},"f:status":{},"f:type":{}},"k:{\"type\":\"FrequentContainerdRestart\"}":{".":{},"f:lastHeartbeatTime":{},"f:lastTransitionTime":{},"f:message":{},"f:reason":{},"f:status":{},"f:type":{}},"k:{\"type\":\"FrequentDockerRestart\"}":{".":{},"f:lastHeartbeatTime":{},"f:lastTransitionTime":{},"f:message":{},"f:reason":{},"f:status":{},"f:type":{}},"k:{\"type\":\"FrequentKubeletRestart\"}":{".":{},"f:lastHeartbeatTime":{},"f:lastTransitionTime":{},"f:message":{},"f:reason":{},"f:status":{},"f:type":{}},"k:{\"type\":\"FrequentUnregisterNetDevice\"}":{".":{},"f:lastHeartbeatTime":{},"f:lastTransitionTime":{},"f:message":{},"f:reason":{},"f:status":{},"f:type":{}},"k:{\"type\":\"KernelDeadlock\"}":{".":{},"f:lastHeartbeatTime":{},"f:lastTransitionTime":{},"f:message":{},"f:reason":{},"f:status":{},"f:type":{}},"k:{\"type\":\"ReadonlyFilesystem\"}":{".":{},"f:lastHeartbeatTime":{},"f:lastTransitionTime":{},"f:message":{},"f:reason":{},"f:status":{},"f:type":{}}}}} status} {kubelet Update v1 2022-11-26 05:27:50 +0000 UTC FieldsV1 {"f:metadata":{"f:labels":{"f:topology.hostpath.csi/node":{}}},"f:status":{"f:conditions":{"k:{\"type\":\"DiskPressure\"}":{"f:lastHeartbeatTime":{}},"k:{\"type\":\"MemoryPressure\"}":{"f:lastHeartbeatTime":{}},"k:{\"type\":\"PIDPressure\"}":{"f:lastHeartbeatTime":{}},"k:{\"type\":\"Ready\"}":{"f:lastHeartbeatTime":{},"f:lastTransitionTime":{},"f:message":{},"f:reason":{},"f:status":{}}},"f:images":{},"f:volumesInUse":{}}} status}]},Spec:NodeSpec{PodCIDR:10.64.0.0/24,DoNotUseExternalID:,ProviderID:gce://k8s-jkns-e2e-gce-ci-reboot/us-west1-b/bootstrap-e2e-minion-group-rwsn,Unschedulable:false,Taints:[]Taint{},ConfigSource:nil,PodCIDRs:[10.64.0.0/24],},Status:NodeStatus{Capacity:ResourceList{attachable-volumes-gce-pd: {{127 0} {<nil>} 127 DecimalSI},cpu: {{2 0} {<nil>} 2 DecimalSI},ephemeral-storage: {{101203873792 0} {<nil>} 98831908Ki BinarySI},hugepages-1Gi: {{0 0} {<nil>} 0 DecimalSI},hugepages-2Mi: {{0 0} {<nil>} 0 DecimalSI},memory: {{7815438336 0} {<nil>} BinarySI},pods: {{110 0} {<nil>} 110 DecimalSI},},Allocatable:ResourceList{attachable-volumes-gce-pd: {{127 0} {<nil>} 127 DecimalSI},cpu: {{2 0} {<nil>} 2 DecimalSI},ephemeral-storage: {{91083486262 0} {<nil>} 91083486262 DecimalSI},hugepages-1Gi: {{0 0} {<nil>} 0 DecimalSI},hugepages-2Mi: {{0 0} {<nil>} 0 DecimalSI},memory: {{7553294336 0} {<nil>} BinarySI},pods: {{110 0} {<nil>} 110 DecimalSI},},Phase:,Conditions:[]NodeCondition{NodeCondition{Type:ReadonlyFilesystem,Status:False,LastHeartbeatTime:2022-11-26 05:25:21 +0000 UTC,LastTransitionTime:2022-11-26 05:05:18 +0000 UTC,Reason:FilesystemIsNotReadOnly,Message:Filesystem is not read-only,},NodeCondition{Type:CorruptDockerOverlay2,Status:False,LastHeartbeatTime:2022-11-26 05:25:21 +0000 UTC,LastTransitionTime:2022-11-26 05:05:18 +0000 UTC,Reason:NoCorruptDockerOverlay2,Message:docker overlay2 is functioning properly,},NodeCondition{Type:FrequentUnregisterNetDevice,Status:False,LastHeartbeatTime:2022-11-26 05:25:21 +0000 UTC,LastTransitionTime:2022-11-26 05:05:18 +0000 UTC,Reason:NoFrequentUnregisterNetDevice,Message:node is functioning properly,},NodeCondition{Type:FrequentKubeletRestart,Status:False,LastHeartbeatTime:2022-11-26 05:25:21 +0000 UTC,LastTransitionTime:2022-11-26 05:05:18 +0000 UTC,Reason:NoFrequentKubeletRestart,Message:kubelet is functioning properly,},NodeCondition{Type:FrequentDockerRestart,Status:False,LastHeartbeatTime:2022-11-26 05:25:21 +0000 UTC,LastTransitionTime:2022-11-26 05:05:18 +0000 UTC,Reason:NoFrequentDockerRestart,Message:docker is functioning properly,},NodeCondition{Type:FrequentContainerdRestart,Status:False,LastHeartbeatTime:2022-11-26 05:25:21 +0000 UTC,LastTransitionTime:2022-11-26 05:05:18 +0000 UTC,Reason:NoFrequentContainerdRestart,Message:containerd is functioning properly,},NodeCondition{Type:KernelDeadlock,Status:False,LastHeartbeatTime:2022-11-26 05:25:21 +0000 UTC,LastTransitionTime:2022-11-26 05:05:18 +0000 UTC,Reason:KernelHasNoDeadlock,Message:kernel has no deadlock,},NodeCondition{Type:NetworkUnavailable,Status:False,LastHeartbeatTime:2022-11-26 05:05:25 +0000 UTC,LastTransitionTime:2022-11-26 05:05:25 +0000 UTC,Reason:RouteCreated,Message:RouteController created a route,},NodeCondition{Type:MemoryPressure,Status:False,LastHeartbeatTime:2022-11-26 05:27:50 +0000 UTC,LastTransitionTime:2022-11-26 05:05:15 +0000 UTC,Reason:KubeletHasSufficientMemory,Message:kubelet has sufficient memory available,},NodeCondition{Type:DiskPressure,Status:False,LastHeartbeatTime:2022-11-26 05:27:50 +0000 UTC,LastTransitionTime:2022-11-26 05:05:15 +0000 UTC,Reason:KubeletHasNoDiskPressure,Message:kubelet has no disk pressure,},NodeCondition{Type:PIDPressure,Status:False,LastHeartbeatTime:2022-11-26 05:27:50 +0000 UTC,LastTransitionTime:2022-11-26 05:05:15 +0000 UTC,Reason:KubeletHasSufficientPID,Message:kubelet has sufficient PID available,},NodeCondition{Type:Ready,Status:True,LastHeartbeatTime:2022-11-26 05:27:50 +0000 UTC,LastTransitionTime:2022-11-26 05:05:17 +0000 UTC,Reason:KubeletReady,Message:kubelet is posting ready status. AppArmor enabled,},},Addresses:[]NodeAddress{NodeAddress{Type:InternalIP,Address:10.138.0.4,},NodeAddress{Type:ExternalIP,Address:35.230.27.45,},NodeAddress{Type:InternalDNS,Address:bootstrap-e2e-minion-group-rwsn.c.k8s-jkns-e2e-gce-ci-reboot.internal,},NodeAddress{Type:Hostname,Address:bootstrap-e2e-minion-group-rwsn.c.k8s-jkns-e2e-gce-ci-reboot.internal,},},DaemonEndpoints:NodeDaemonEndpoints{KubeletEndpoint:DaemonEndpoint{Port:10250,},},NodeInfo:NodeSystemInfo{MachineID:69e22ce1281686d6cde1b3211b34b8af,SystemUUID:69e22ce1-2816-86d6-cde1-b3211b34b8af,BootID:79e9e1fd-2a39-42ec-b72b-72836d213862,KernelVersion:5.10.123+,OSImage:Container-Optimized OS from Google,ContainerRuntimeVersion:containerd://1.7.0-beta.0-149-gd06318622,KubeletVersion:v1.27.0-alpha.0.50+70617042976dc1,KubeProxyVersion:v1.27.0-alpha.0.50+70617042976dc1,OperatingSystem:linux,Architecture:amd64,},Images:[]ContainerImage{ContainerImage{Names:[registry.k8s.io/e2e-test-images/jessie-dnsutils@sha256:24aaf2626d6b27864c29de2097e8bbb840b3a414271bf7c8995e431e47d8408e registry.k8s.io/e2e-test-images/jessie-dnsutils:1.7],SizeBytes:112030336,},ContainerImage{Names:[registry.k8s.io/e2e-test-images/volume/nfs@sha256:3bda73f2428522b0e342af80a0b9679e8594c2126f2b3cca39ed787589741b9e registry.k8s.io/e2e-test-images/volume/nfs:1.3],SizeBytes:95836203,},ContainerImage{Names:[registry.k8s.io/sig-storage/nfs-provisioner@sha256:e943bb77c7df05ebdc8c7888b2db289b13bf9f012d6a3a5a74f14d4d5743d439 registry.k8s.io/sig-storage/nfs-provisioner:v3.0.1],SizeBytes:90632047,},ContainerImage{Names:[registry.k8s.io/kube-proxy-amd64:v1.27.0-alpha.0.50_70617042976dc1],SizeBytes:67201736,},ContainerImage{Names:[registry.k8s.io/e2e-test-images/agnhost@sha256:16bbf38c463a4223d8cfe4da12bc61010b082a79b4bb003e2d3ba3ece5dd5f9e registry.k8s.io/e2e-test-images/agnhost:2.43],SizeBytes:51706353,},ContainerImage{Names:[gke.gcr.io/prometheus-to-sd@sha256:e739643c3939ba0b161425f45a1989eedfc4a3b166db9a7100863296b4c70510 gke.gcr.io/prometheus-to-sd:v0.11.1-gke.1],SizeBytes:48742566,},ContainerImage{Names:[registry.k8s.io/metrics-server/metrics-server@sha256:6385aec64bb97040a5e692947107b81e178555c7a5b71caa90d733e4130efc10 registry.k8s.io/metrics-server/metrics-server:v0.5.2],SizeBytes:26023008,},ContainerImage{Names:[registry.k8s.io/sig-storage/csi-provisioner@sha256:ee3b525d5b89db99da3b8eb521d9cd90cb6e9ef0fbb651e98bb37be78d36b5b8 registry.k8s.io/sig-storage/csi-provisioner:v3.3.0],SizeBytes:25491225,},ContainerImage{Names:[registry.k8s.io/sig-storage/csi-resizer@sha256:425d8f1b769398127767b06ed97ce62578a3179bcb99809ce93a1649e025ffe7 registry.k8s.io/sig-storage/csi-resizer:v1.6.0],SizeBytes:24148884,},ContainerImage{Names:[registry.k8s.io/sig-storage/csi-snapshotter@sha256:291334908ddf71a4661fd7f6d9d97274de8a5378a2b6fdfeb2ce73414a34f82f registry.k8s.io/sig-storage/csi-snapshotter:v6.1.0],SizeBytes:23881995,},ContainerImage{Names:[registry.k8s.io/sig-storage/csi-attacher@sha256:9a685020911e2725ad019dbce6e4a5ab93d51e3d4557f115e64343345e05781b registry.k8s.io/sig-storage/csi-attacher:v4.0.0],SizeBytes:23847201,},ContainerImage{Names:[registry.k8s.io/sig-storage/snapshot-controller@sha256:823c75d0c45d1427f6d850070956d9ca657140a7bbf828381541d1d808475280 registry.k8s.io/sig-storage/snapshot-controller:v6.1.0],SizeBytes:22620891,},ContainerImage{Names:[registry.k8s.io/sig-storage/hostpathplugin@sha256:92257881c1d6493cf18299a24af42330f891166560047902b8d431fb66b01af5 registry.k8s.io/sig-storage/hostpathplugin:v1.9.0],SizeBytes:18758628,},ContainerImage{Names:[registry.k8s.io/cpa/cluster-proportional-autoscaler@sha256:fd636b33485c7826fb20ef0688a83ee0910317dbb6c0c6f3ad14661c1db25def registry.k8s.io/cpa/cluster-proportional-autoscaler:1.8.4],SizeBytes:15209393,},ContainerImage{Names:[registry.k8s.io/coredns/coredns@sha256:8e352a029d304ca7431c6507b56800636c321cb52289686a581ab70aaa8a2e2a registry.k8s.io/coredns/coredns:v1.9.3],SizeBytes:14837849,},ContainerImage{Names:[registry.k8s.io/autoscaling/addon-resizer@sha256:43f129b81d28f0fdd54de6d8e7eacd5728030782e03db16087fc241ad747d3d6 registry.k8s.io/autoscaling/addon-resizer:1.8.14],SizeBytes:10153852,},ContainerImage{Names:[registry.k8s.io/sig-storage/csi-node-driver-registrar@sha256:0103eee7c35e3e0b5cd8cdca9850dc71c793cdeb6669d8be7a89440da2d06ae4 registry.k8s.io/sig-storage/csi-node-driver-registrar:v2.5.1],SizeBytes:9133109,},ContainerImage{Names:[registry.k8s.io/sig-storage/livenessprobe@sha256:933940f13b3ea0abc62e656c1aa5c5b47c04b15d71250413a6b821bd0c58b94e registry.k8s.io/sig-storage/livenessprobe:v2.7.0],SizeBytes:8688564,},ContainerImage{Names:[registry.k8s.io/kas-network-proxy/proxy-agent@sha256:48f2a4ec3e10553a81b8dd1c6fa5fe4bcc9617f78e71c1ca89c6921335e2d7da registry.k8s.io/kas-network-proxy/proxy-agent:v0.0.33],SizeBytes:8512162,},ContainerImage{Names:[registry.k8s.io/networking/ingress-gce-404-server-with-metrics-amd64@sha256:7eb7b3cee4d33c10c49893ad3c386232b86d4067de5251294d4c620d6e072b93 registry.k8s.io/networking/ingress-gce-404-server-with-metrics-amd64:v1.10.11],SizeBytes:6463068,},ContainerImage{Names:[registry.k8s.io/metadata-proxy@sha256:e914645f22e946bce5165737e1b244e0a296ad1f0f81a9531adc57af2780978a registry.k8s.io/metadata-proxy:v0.1.12],SizeBytes:5301657,},ContainerImage{Names:[registry.k8s.io/e2e-test-images/busybox@sha256:c318242786b139d18676b1c09a0ad7f15fc17f8f16a5b2e625cd0dc8c9703daf registry.k8s.io/e2e-test-images/busybox:1.29-2],SizeBytes:732424,},ContainerImage{Names:[registry.k8s.io/e2e-test-images/busybox@sha256:2e0f836850e09b8b7cc937681d6194537a09fbd5f6b9e08f4d646a85128e8937 registry.k8s.io/e2e-test-images/busybox:1.29-4],SizeBytes:731990,},ContainerImage{Names:[registry.k8s.io/pause@sha256:7031c1b283388d2c2e09b57badb803c05ebed362dc88d84b480cc47f72a21097 registry.k8s.io/pause:3.9],SizeBytes:321520,},ContainerImage{Names:[registry.k8s.io/pause@sha256:9001185023633d17a2f98ff69b6ff2615b8ea02a825adffa40422f51dfdcde9d registry.k8s.io/pause:3.8],SizeBytes:311286,},},VolumesInUse:[kubernetes.io/csi/csi-hostpath-multivolume-4683^6fe7a3e3-6d49-11ed-ac34-8a6e303c97ed kubernetes.io/csi/csi-hostpath-multivolume-4683^7ad9a9cd-6d49-11ed-b2a0-f2f9c395d3b2 kubernetes.io/csi/csi-hostpath-multivolume-6872^8dfb6ea6-6d49-11ed-bbbd-7a3847645704 kubernetes.io/csi/csi-hostpath-provisioning-6984^3b0457df-6d48-11ed-9e4d-d626cdd9a50a],VolumesAttached:[]AttachedVolume{AttachedVolume{Name:kubernetes.io/csi/csi-hostpath-provisioning-6984^3b0457df-6d48-11ed-9e4d-d626cdd9a50a,DevicePath:,},AttachedVolume{Name:kubernetes.io/csi/csi-hostpath-multivolume-4683^7ad9a9cd-6d49-11ed-b2a0-f2f9c395d3b2,DevicePath:,},AttachedVolume{Name:kubernetes.io/csi/csi-hostpath-multivolume-4683^6fe7a3e3-6d49-11ed-ac34-8a6e303c97ed,DevicePath:,},AttachedVolume{Name:kubernetes.io/csi/csi-hostpath-multivolume-6872^8dfb6ea6-6d49-11ed-bbbd-7a3847645704,DevicePath:,},AttachedVolume{Name:kubernetes.io/csi/csi-hostpath-provisioning-8014^72c629d1-6d4a-11ed-a786-1af3032da8b9,DevicePath:,},},Config:nil,},} Nov 26 05:27:52.249: INFO: Logging kubelet events for node bootstrap-e2e-minion-group-rwsn Nov 26 05:27:52.293: INFO: Logging pods the kubelet thinks is on node bootstrap-e2e-minion-group-rwsn Nov 26 05:27:52.357: INFO: kube-dns-autoscaler-5f6455f985-xktw8 started at 2022-11-26 05:05:25 +0000 UTC (0+1 container statuses recorded) Nov 26 05:27:52.357: INFO: Container autoscaler ready: false, restart count 7 Nov 26 05:27:52.357: INFO: coredns-6d97d5ddb-p2xpt started at 2022-11-26 05:05:31 +0000 UTC (0+1 container statuses recorded) Nov 26 05:27:52.357: INFO: Container coredns ready: false, restart count 8 Nov 26 05:27:52.357: INFO: pod-04078049-bcc4-43bf-8abc-addcb96042db started at 2022-11-26 05:16:29 +0000 UTC (0+1 container statuses recorded) Nov 26 05:27:52.357: INFO: Container write-pod ready: false, restart count 0 Nov 26 05:27:52.357: INFO: kube-proxy-bootstrap-e2e-minion-group-rwsn started at 2022-11-26 05:05:15 +0000 UTC (0+1 container statuses recorded) Nov 26 05:27:52.357: INFO: Container kube-proxy ready: false, restart count 7 Nov 26 05:27:52.357: INFO: csi-hostpathplugin-0 started at 2022-11-26 05:23:21 +0000 UTC (0+7 container statuses recorded) Nov 26 05:27:52.357: INFO: Container csi-attacher ready: false, restart count 5 Nov 26 05:27:52.357: INFO: Container csi-provisioner ready: false, restart count 5 Nov 26 05:27:52.357: INFO: Container csi-resizer ready: false, restart count 5 Nov 26 05:27:52.357: INFO: Container csi-snapshotter ready: false, restart count 5 Nov 26 05:27:52.357: INFO: Container hostpath ready: false, restart count 5 Nov 26 05:27:52.357: INFO: Container liveness-probe ready: false, restart count 5 Nov 26 05:27:52.357: INFO: Container node-driver-registrar ready: false, restart count 5 Nov 26 05:27:52.357: INFO: netserver-1 started at 2022-11-26 05:11:06 +0000 UTC (0+1 container statuses recorded) Nov 26 05:27:52.357: INFO: Container webserver ready: true, restart count 7 Nov 26 05:27:52.357: INFO: pvc-volume-tester-5h2gx started at 2022-11-26 05:13:54 +0000 UTC (0+1 container statuses recorded) Nov 26 05:27:52.357: INFO: Container volume-tester ready: false, restart count 0 Nov 26 05:27:52.357: INFO: metadata-proxy-v0.1-qlckk started at 2022-11-26 05:05:16 +0000 UTC (0+2 container statuses recorded) Nov 26 05:27:52.357: INFO: Container metadata-proxy ready: true, restart count 0 Nov 26 05:27:52.357: INFO: Container prometheus-to-sd-exporter ready: true, restart count 0 Nov 26 05:27:52.357: INFO: csi-hostpathplugin-0 started at 2022-11-26 05:10:11 +0000 UTC (0+7 container statuses recorded) Nov 26 05:27:52.357: INFO: Container csi-attacher ready: false, restart count 7 Nov 26 05:27:52.357: INFO: Container csi-provisioner ready: false, restart count 7 Nov 26 05:27:52.357: INFO: Container csi-resizer ready: false, restart count 7 Nov 26 05:27:52.357: INFO: Container csi-snapshotter ready: false, restart count 7 Nov 26 05:27:52.357: INFO: Container hostpath ready: false, restart count 7 Nov 26 05:27:52.357: INFO: Container liveness-probe ready: false, restart count 7 Nov 26 05:27:52.357: INFO: Container node-driver-registrar ready: false, restart count 8 Nov 26 05:27:52.357: INFO: csi-hostpathplugin-0 started at 2022-11-26 05:16:04 +0000 UTC (0+7 container statuses recorded) Nov 26 05:27:52.357: INFO: Container csi-attacher ready: false, restart count 5 Nov 26 05:27:52.357: INFO: Container csi-provisioner ready: false, restart count 5 Nov 26 05:27:52.357: INFO: Container csi-resizer ready: false, restart count 5 Nov 26 05:27:52.357: INFO: Container csi-snapshotter ready: false, restart count 5 Nov 26 05:27:52.357: INFO: Container hostpath ready: false, restart count 5 Nov 26 05:27:52.357: INFO: Container liveness-probe ready: false, restart count 5 Nov 26 05:27:52.357: INFO: Container node-driver-registrar ready: false, restart count 5 Nov 26 05:27:52.357: INFO: konnectivity-agent-sxg5q started at 2022-11-26 05:05:25 +0000 UTC (0+1 container statuses recorded) Nov 26 05:27:52.357: INFO: Container konnectivity-agent ready: false, restart count 7 Nov 26 05:27:52.357: INFO: pod-ce08e819-7446-4933-9e24-3a40ac81db8b started at 2022-11-26 05:17:01 +0000 UTC (0+1 container statuses recorded) Nov 26 05:27:52.357: INFO: Container write-pod ready: false, restart count 0 Nov 26 05:27:52.357: INFO: csi-mockplugin-0 started at 2022-11-26 05:23:19 +0000 UTC (0+4 container statuses recorded) Nov 26 05:27:52.357: INFO: Container busybox ready: false, restart count 3 Nov 26 05:27:52.357: INFO: Container csi-provisioner ready: true, restart count 3 Nov 26 05:27:52.357: INFO: Container driver-registrar ready: true, restart count 4 Nov 26 05:27:52.357: INFO: Container mock ready: true, restart count 4 Nov 26 05:27:52.357: INFO: csi-mockplugin-0 started at 2022-11-26 05:13:01 +0000 UTC (0+4 container statuses recorded) Nov 26 05:27:52.357: INFO: Container busybox ready: false, restart count 5 Nov 26 05:27:52.357: INFO: Container csi-provisioner ready: false, restart count 5 Nov 26 05:27:52.357: INFO: Container driver-registrar ready: false, restart count 6 Nov 26 05:27:52.357: INFO: Container mock ready: false, restart count 6 Nov 26 05:27:52.357: INFO: hostexec-bootstrap-e2e-minion-group-rwsn-l47jh started at 2022-11-26 05:19:38 +0000 UTC (0+1 container statuses recorded) Nov 26 05:27:52.357: INFO: Container agnhost-container ready: false, restart count 4 Nov 26 05:27:52.357: INFO: affinity-lb-esipp-4hthn started at 2022-11-26 05:23:40 +0000 UTC (0+1 container statuses recorded) Nov 26 05:27:52.357: INFO: Container affinity-lb-esipp ready: true, restart count 2 Nov 26 05:27:52.357: INFO: l7-default-backend-8549d69d99-8cfzv started at 2022-11-26 05:05:25 +0000 UTC (0+1 container statuses recorded) Nov 26 05:27:52.357: INFO: Container default-http-backend ready: true, restart count 0 Nov 26 05:27:52.357: INFO: volume-snapshot-controller-0 started at 2022-11-26 05:05:25 +0000 UTC (0+1 container statuses recorded) Nov 26 05:27:52.357: INFO: Container volume-snapshot-controller ready: true, restart count 6 Nov 26 05:27:52.357: INFO: csi-mockplugin-0 started at 2022-11-26 05:08:11 +0000 UTC (0+4 container statuses recorded) Nov 26 05:27:52.357: INFO: Container busybox ready: false, restart count 5 Nov 26 05:27:52.357: INFO: Container csi-provisioner ready: false, restart count 7 Nov 26 05:27:52.357: INFO: Container driver-registrar ready: false, restart count 5 Nov 26 05:27:52.357: INFO: Container mock ready: false, restart count 5 Nov 26 05:27:52.357: INFO: netserver-1 started at 2022-11-26 05:11:29 +0000 UTC (0+1 container statuses recorded) Nov 26 05:27:52.357: INFO: Container webserver ready: true, restart count 5 Nov 26 05:27:52.357: INFO: netserver-1 started at 2022-11-26 05:09:51 +0000 UTC (0+1 container statuses recorded) Nov 26 05:27:52.357: INFO: Container webserver ready: false, restart count 6 Nov 26 05:27:52.357: INFO: hostexec-bootstrap-e2e-minion-group-rwsn-l7nb2 started at 2022-11-26 05:23:35 +0000 UTC (0+1 container statuses recorded) Nov 26 05:27:52.357: INFO: Container agnhost-container ready: true, restart count 2 Nov 26 05:27:52.357: INFO: csi-hostpathplugin-0 started at 2022-11-26 05:16:00 +0000 UTC (0+7 container statuses recorded) Nov 26 05:27:52.357: INFO: Container csi-attacher ready: false, restart count 6 Nov 26 05:27:52.357: INFO: Container csi-provisioner ready: false, restart count 6 Nov 26 05:27:52.357: INFO: Container csi-resizer ready: false, restart count 6 Nov 26 05:27:52.357: INFO: Container csi-snapshotter ready: false, restart count 6 Nov 26 05:27:52.357: INFO: Container hostpath ready: false, restart count 6 Nov 26 05:27:52.357: INFO: Container liveness-probe ready: false, restart count 6 Nov 26 05:27:52.357: INFO: Container node-driver-registrar ready: false, restart count 6 Nov 26 05:27:52.357: INFO: coredns-6d97d5ddb-xk6v4 started at 2022-11-26 05:05:25 +0000 UTC (0+1 container statuses recorded) Nov 26 05:27:52.357: INFO: Container coredns ready: false, restart count 8 Nov 26 05:27:52.357: INFO: pod-subpath-test-dynamicpv-h6m7 started at 2022-11-26 05:07:31 +0000 UTC (1+1 container statuses recorded) Nov 26 05:27:52.357: INFO: Init container init-volume-dynamicpv-h6m7 ready: true, restart count 0 Nov 26 05:27:52.357: INFO: Container test-container-subpath-dynamicpv-h6m7 ready: false, restart count 0 Nov 26 05:27:52.357: INFO: hostexec-bootstrap-e2e-minion-group-rwsn-8s6qn started at 2022-11-26 05:23:14 +0000 UTC (0+1 container statuses recorded) Nov 26 05:27:52.357: INFO: Container agnhost-container ready: true, restart count 3 Nov 26 05:27:52.357: INFO: hostexec-bootstrap-e2e-minion-group-rwsn-ssq5n started at 2022-11-26 05:23:33 +0000 UTC (0+1 container statuses recorded) Nov 26 05:27:52.357: INFO: Container agnhost-container ready: true, restart count 0 Nov 26 05:27:52.574: INFO: Latency metrics for node bootstrap-e2e-minion-group-rwsn Nov 26 05:27:52.574: INFO: Logging node info for node bootstrap-e2e-minion-group-vhdj Nov 26 05:27:52.616: INFO: Node Info: &Node{ObjectMeta:{bootstrap-e2e-minion-group-vhdj 0d27fd16-9ec8-4428-bb1e-97fead6c7a66 10875 0 2022-11-26 05:05:23 +0000 UTC <nil> <nil> map[beta.kubernetes.io/arch:amd64 beta.kubernetes.io/instance-type:n1-standard-2 beta.kubernetes.io/os:linux cloud.google.com/metadata-proxy-ready:true failure-domain.beta.kubernetes.io/region:us-west1 failure-domain.beta.kubernetes.io/zone:us-west1-b kubernetes.io/arch:amd64 kubernetes.io/hostname:bootstrap-e2e-minion-group-vhdj kubernetes.io/os:linux node.kubernetes.io/instance-type:n1-standard-2 topology.hostpath.csi/node:bootstrap-e2e-minion-group-vhdj topology.kubernetes.io/region:us-west1 topology.kubernetes.io/zone:us-west1-b] map[csi.volume.kubernetes.io/nodeid:{"csi-hostpath-provisioning-7617":"bootstrap-e2e-minion-group-vhdj","csi-hostpath-volumeio-4786":"bootstrap-e2e-minion-group-vhdj"} node.alpha.kubernetes.io/ttl:0 volumes.kubernetes.io/controller-managed-attach-detach:true] [] [] [{kubelet Update v1 2022-11-26 05:05:23 +0000 UTC FieldsV1 {"f:metadata":{"f:annotations":{".":{},"f:volumes.kubernetes.io/controller-managed-attach-detach":{}},"f:labels":{".":{},"f:beta.kubernetes.io/arch":{},"f:beta.kubernetes.io/instance-type":{},"f:beta.kubernetes.io/os":{},"f:cloud.google.com/metadata-proxy-ready":{},"f:failure-domain.beta.kubernetes.io/region":{},"f:failure-domain.beta.kubernetes.io/zone":{},"f:kubernetes.io/arch":{},"f:kubernetes.io/hostname":{},"f:kubernetes.io/os":{},"f:node.kubernetes.io/instance-type":{},"f:topology.kubernetes.io/region":{},"f:topology.kubernetes.io/zone":{}}},"f:spec":{"f:providerID":{}}} } {kube-controller-manager Update v1 2022-11-26 05:05:24 +0000 UTC FieldsV1 {"f:metadata":{"f:annotations":{"f:node.alpha.kubernetes.io/ttl":{}}},"f:spec":{"f:podCIDR":{},"f:podCIDRs":{".":{},"v:\"10.64.3.0/24\"":{}}}} } {kube-controller-manager Update v1 2022-11-26 05:17:38 +0000 UTC FieldsV1 {"f:status":{"f:conditions":{"k:{\"type\":\"NetworkUnavailable\"}":{"f:lastHeartbeatTime":{},"f:lastTransitionTime":{},"f:message":{},"f:reason":{},"f:status":{}}}}} status} {node-problem-detector Update v1 2022-11-26 05:25:29 +0000 UTC FieldsV1 {"f:status":{"f:conditions":{"k:{\"type\":\"CorruptDockerOverlay2\"}":{".":{},"f:lastHeartbeatTime":{},"f:lastTransitionTime":{},"f:message":{},"f:reason":{},"f:status":{},"f:type":{}},"k:{\"type\":\"FrequentContainerdRestart\"}":{".":{},"f:lastHeartbeatTime":{},"f:lastTransitionTime":{},"f:message":{},"f:reason":{},"f:status":{},"f:type":{}},"k:{\"type\":\"FrequentDockerRestart\"}":{".":{},"f:lastHeartbeatTime":{},"f:lastTransitionTime":{},"f:message":{},"f:reason":{},"f:status":{},"f:type":{}},"k:{\"type\":\"FrequentKubeletRestart\"}":{".":{},"f:lastHeartbeatTime":{},"f:lastTransitionTime":{},"f:message":{},"f:reason":{},"f:status":{},"f:type":{}},"k:{\"type\":\"FrequentUnregisterNetDevice\"}":{".":{},"f:lastHeartbeatTime":{},"f:lastTransitionTime":{},"f:message":{},"f:reason":{},"f:status":{},"f:type":{}},"k:{\"type\":\"KernelDeadlock\"}":{".":{},"f:lastHeartbeatTime":{},"f:lastTransitionTime":{},"f:message":{},"f:reason":{},"f:status":{},"f:type":{}},"k:{\"type\":\"ReadonlyFilesystem\"}":{".":{},"f:lastHeartbeatTime":{},"f:lastTransitionTime":{},"f:message":{},"f:reason":{},"f:status":{},"f:type":{}}}}} status} {kubelet Update v1 2022-11-26 05:27:51 +0000 UTC FieldsV1 {"f:metadata":{"f:annotations":{"f:csi.volume.kubernetes.io/nodeid":{}},"f:labels":{"f:topology.hostpath.csi/node":{}}},"f:status":{"f:conditions":{"k:{\"type\":\"DiskPressure\"}":{"f:lastHeartbeatTime":{}},"k:{\"type\":\"MemoryPressure\"}":{"f:lastHeartbeatTime":{}},"k:{\"type\":\"PIDPressure\"}":{"f:lastHeartbeatTime":{}},"k:{\"type\":\"Ready\"}":{"f:lastHeartbeatTime":{},"f:lastTransitionTime":{},"f:message":{},"f:reason":{},"f:status":{}}},"f:images":{}}} status}]},Spec:NodeSpec{PodCIDR:10.64.3.0/24,DoNotUseExternalID:,ProviderID:gce://k8s-jkns-e2e-gce-ci-reboot/us-west1-b/bootstrap-e2e-minion-group-vhdj,Unschedulable:false,Taints:[]Taint{},ConfigSource:nil,PodCIDRs:[10.64.3.0/24],},Status:NodeStatus{Capacity:ResourceList{attachable-volumes-gce-pd: {{127 0} {<nil>} 127 DecimalSI},cpu: {{2 0} {<nil>} 2 DecimalSI},ephemeral-storage: {{101203873792 0} {<nil>} 98831908Ki BinarySI},hugepages-1Gi: {{0 0} {<nil>} 0 DecimalSI},hugepages-2Mi: {{0 0} {<nil>} 0 DecimalSI},memory: {{7815438336 0} {<nil>} BinarySI},pods: {{110 0} {<nil>} 110 DecimalSI},},Allocatable:ResourceList{attachable-volumes-gce-pd: {{127 0} {<nil>} 127 DecimalSI},cpu: {{2 0} {<nil>} 2 DecimalSI},ephemeral-storage: {{91083486262 0} {<nil>} 91083486262 DecimalSI},hugepages-1Gi: {{0 0} {<nil>} 0 DecimalSI},hugepages-2Mi: {{0 0} {<nil>} 0 DecimalSI},memory: {{7553294336 0} {<nil>} BinarySI},pods: {{110 0} {<nil>} 110 DecimalSI},},Phase:,Conditions:[]NodeCondition{NodeCondition{Type:FrequentKubeletRestart,Status:False,LastHeartbeatTime:2022-11-26 05:25:29 +0000 UTC,LastTransitionTime:2022-11-26 05:05:26 +0000 UTC,Reason:NoFrequentKubeletRestart,Message:kubelet is functioning properly,},NodeCondition{Type:FrequentDockerRestart,Status:False,LastHeartbeatTime:2022-11-26 05:25:29 +0000 UTC,LastTransitionTime:2022-11-26 05:05:26 +0000 UTC,Reason:NoFrequentDockerRestart,Message:docker is functioning properly,},NodeCondition{Type:FrequentContainerdRestart,Status:False,LastHeartbeatTime:2022-11-26 05:25:29 +0000 UTC,LastTransitionTime:2022-11-26 05:05:26 +0000 UTC,Reason:NoFrequentContainerdRestart,Message:containerd is functioning properly,},NodeCondition{Type:KernelDeadlock,Status:False,LastHeartbeatTime:2022-11-26 05:25:29 +0000 UTC,LastTransitionTime:2022-11-26 05:05:26 +0000 UTC,Reason:KernelHasNoDeadlock,Message:kernel has no deadlock,},NodeCondition{Type:ReadonlyFilesystem,Status:False,LastHeartbeatTime:2022-11-26 05:25:29 +0000 UTC,LastTransitionTime:2022-11-26 05:05:26 +0000 UTC,Reason:FilesystemIsNotReadOnly,Message:Filesystem is not read-only,},NodeCondition{Type:CorruptDockerOverlay2,Status:False,LastHeartbeatTime:2022-11-26 05:25:29 +0000 UTC,LastTransitionTime:2022-11-26 05:05:26 +0000 UTC,Reason:NoCorruptDockerOverlay2,Message:docker overlay2 is functioning properly,},NodeCondition{Type:FrequentUnregisterNetDevice,Status:False,LastHeartbeatTime:2022-11-26 05:25:29 +0000 UTC,LastTransitionTime:2022-11-26 05:05:26 +0000 UTC,Reason:NoFrequentUnregisterNetDevice,Message:node is functioning properly,},NodeCondition{Type:NetworkUnavailable,Status:False,LastHeartbeatTime:2022-11-26 05:05:39 +0000 UTC,LastTransitionTime:2022-11-26 05:05:39 +0000 UTC,Reason:RouteCreated,Message:RouteController created a route,},NodeCondition{Type:MemoryPressure,Status:False,LastHeartbeatTime:2022-11-26 05:27:51 +0000 UTC,LastTransitionTime:2022-11-26 05:05:23 +0000 UTC,Reason:KubeletHasSufficientMemory,Message:kubelet has sufficient memory available,},NodeCondition{Type:DiskPressure,Status:False,LastHeartbeatTime:2022-11-26 05:27:51 +0000 UTC,LastTransitionTime:2022-11-26 05:05:23 +0000 UTC,Reason:KubeletHasNoDiskPressure,Message:kubelet has no disk pressure,},NodeCondition{Type:PIDPressure,Status:False,LastHeartbeatTime:2022-11-26 05:27:51 +0000 UTC,LastTransitionTime:2022-11-26 05:05:23 +0000 UTC,Reason:KubeletHasSufficientPID,Message:kubelet has sufficient PID available,},NodeCondition{Type:Ready,Status:True,LastHeartbeatTime:2022-11-26 05:27:51 +0000 UTC,LastTransitionTime:2022-11-26 05:05:24 +0000 UTC,Reason:KubeletReady,Message:kubelet is posting ready status. AppArmor enabled,},},Addresses:[]NodeAddress{NodeAddress{Type:InternalIP,Address:10.138.0.3,},NodeAddress{Type:ExternalIP,Address:34.83.94.215,},NodeAddress{Type:InternalDNS,Address:bootstrap-e2e-minion-group-vhdj.c.k8s-jkns-e2e-gce-ci-reboot.internal,},NodeAddress{Type:Hostname,Address:bootstrap-e2e-minion-group-vhdj.c.k8s-jkns-e2e-gce-ci-reboot.internal,},},DaemonEndpoints:NodeDaemonEndpoints{KubeletEndpoint:DaemonEndpoint{Port:10250,},},NodeInfo:NodeSystemInfo{MachineID:31b0953ad3e740aeb7fc1a89edb5435a,SystemUUID:31b0953a-d3e7-40ae-b7fc-1a89edb5435a,BootID:14c054d0-bda9-4ddf-b9f6-e1742be33d25,KernelVersion:5.10.123+,OSImage:Container-Optimized OS from Google,ContainerRuntimeVersion:containerd://1.7.0-beta.0-149-gd06318622,KubeletVersion:v1.27.0-alpha.0.50+70617042976dc1,KubeProxyVersion:v1.27.0-alpha.0.50+70617042976dc1,OperatingSystem:linux,Architecture:amd64,},Images:[]ContainerImage{ContainerImage{Names:[registry.k8s.io/e2e-test-images/jessie-dnsutils@sha256:24aaf2626d6b27864c29de2097e8bbb840b3a414271bf7c8995e431e47d8408e registry.k8s.io/e2e-test-images/jessie-dnsutils:1.7],SizeBytes:112030336,},ContainerImage{Names:[registry.k8s.io/sig-storage/nfs-provisioner@sha256:e943bb77c7df05ebdc8c7888b2db289b13bf9f012d6a3a5a74f14d4d5743d439 registry.k8s.io/sig-storage/nfs-provisioner:v3.0.1],SizeBytes:90632047,},ContainerImage{Names:[registry.k8s.io/kube-proxy-amd64:v1.27.0-alpha.0.50_70617042976dc1],SizeBytes:67201736,},ContainerImage{Names:[registry.k8s.io/e2e-test-images/agnhost@sha256:16bbf38c463a4223d8cfe4da12bc61010b082a79b4bb003e2d3ba3ece5dd5f9e registry.k8s.io/e2e-test-images/agnhost:2.43],SizeBytes:51706353,},ContainerImage{Names:[gke.gcr.io/prometheus-to-sd@sha256:e739643c3939ba0b161425f45a1989eedfc4a3b166db9a7100863296b4c70510 gke.gcr.io/prometheus-to-sd:v0.11.1-gke.1],SizeBytes:48742566,},ContainerImage{Names:[registry.k8s.io/e2e-test-images/httpd@sha256:148b022f5c5da426fc2f3c14b5c0867e58ef05961510c84749ac1fddcb0fef22 registry.k8s.io/e2e-test-images/httpd:2.4.38-4],SizeBytes:40764257,},ContainerImage{Names:[registry.k8s.io/sig-storage/csi-provisioner@sha256:ee3b525d5b89db99da3b8eb521d9cd90cb6e9ef0fbb651e98bb37be78d36b5b8 registry.k8s.io/sig-storage/csi-provisioner:v3.3.0],SizeBytes:25491225,},ContainerImage{Names:[registry.k8s.io/sig-storage/csi-resizer@sha256:425d8f1b769398127767b06ed97ce62578a3179bcb99809ce93a1649e025ffe7 registry.k8s.io/sig-storage/csi-resizer:v1.6.0],SizeBytes:24148884,},ContainerImage{Names:[registry.k8s.io/sig-storage/csi-snapshotter@sha256:291334908ddf71a4661fd7f6d9d97274de8a5378a2b6fdfeb2ce73414a34f82f registry.k8s.io/sig-storage/csi-snapshotter:v6.1.0],SizeBytes:23881995,},ContainerImage{Names:[registry.k8s.io/sig-storage/csi-attacher@sha256:9a685020911e2725ad019dbce6e4a5ab93d51e3d4557f115e64343345e05781b registry.k8s.io/sig-storage/csi-attacher:v4.0.0],SizeBytes:23847201,},ContainerImage{Names:[registry.k8s.io/sig-storage/hostpathplugin@sha256:92257881c1d6493cf18299a24af42330f891166560047902b8d431fb66b01af5 registry.k8s.io/sig-storage/hostpathplugin:v1.9.0],SizeBytes:18758628,},ContainerImage{Names:[registry.k8s.io/sig-storage/csi-node-driver-registrar@sha256:0103eee7c35e3e0b5cd8cdca9850dc71c793cdeb6669d8be7a89440da2d06ae4 registry.k8s.io/sig-storage/csi-node-driver-registrar:v2.5.1],SizeBytes:9133109,},ContainerImage{Names:[registry.k8s.io/sig-storage/livenessprobe@sha256:933940f13b3ea0abc62e656c1aa5c5b47c04b15d71250413a6b821bd0c58b94e registry.k8s.io/sig-storage/livenessprobe:v2.7.0],SizeBytes:8688564,},ContainerImage{Names:[registry.k8s.io/kas-network-proxy/proxy-agent@sha256:48f2a4ec3e10553a81b8dd1c6fa5fe4bcc9617f78e71c1ca89c6921335e2d7da registry.k8s.io/kas-network-proxy/proxy-agent:v0.0.33],SizeBytes:8512162,},ContainerImage{Names:[registry.k8s.io/metadata-proxy@sha256:e914645f22e946bce5165737e1b244e0a296ad1f0f81a9531adc57af2780978a registry.k8s.io/metadata-proxy:v0.1.12],SizeBytes:5301657,},ContainerImage{Names:[registry.k8s.io/e2e-test-images/busybox@sha256:2e0f836850e09b8b7cc937681d6194537a09fbd5f6b9e08f4d646a85128e8937 registry.k8s.io/e2e-test-images/busybox:1.29-4],SizeBytes:731990,},ContainerImage{Names:[registry.k8s.io/pause@sha256:9001185023633d17a2f98ff69b6ff2615b8ea02a825adffa40422f51dfdcde9d registry.k8s.io/pause:3.8],SizeBytes:311286,},},VolumesInUse:[],VolumesAttached:[]AttachedVolume{},Config:nil,},} Nov 26 05:27:52.616: INFO: Logging kubelet events for node bootstrap-e2e-minion-group-vhdj Nov 26 05:27:52.660: INFO: Logging pods the kubelet thinks is on node bootstrap-e2e-minion-group-vhdj Nov 26 05:27:52.726: INFO: kube-proxy-bootstrap-e2e-minion-group-vhdj started at 2022-11-26 05:05:23 +0000 UTC (0+1 container statuses recorded) Nov 26 05:27:52.726: INFO: Container kube-proxy ready: true, restart count 8 Nov 26 05:27:52.726: INFO: pod-9fe7e984-9a6a-4bb0-9c65-0628a6a2f792 started at 2022-11-26 05:08:13 +0000 UTC (0+1 container statuses recorded) Nov 26 05:27:52.726: INFO: Container write-pod ready: false, restart count 0 Nov 26 05:27:52.726: INFO: metadata-proxy-v0.1-wn5l5 started at 2022-11-26 05:05:24 +0000 UTC (0+2 container statuses recorded) Nov 26 05:27:52.726: INFO: Container metadata-proxy ready: true, restart count 0 Nov 26 05:27:52.726: INFO: Container prometheus-to-sd-exporter ready: true, restart count 0 Nov 26 05:27:52.726: INFO: pod-subpath-test-dynamicpv-4nfp started at 2022-11-26 05:11:17 +0000 UTC (1+2 container statuses recorded) Nov 26 05:27:52.726: INFO: Init container init-volume-dynamicpv-4nfp ready: true, restart count 0 Nov 26 05:27:52.726: INFO: Container test-container-subpath-dynamicpv-4nfp ready: false, restart count 2 Nov 26 05:27:52.726: INFO: Container test-container-volume-dynamicpv-4nfp ready: false, restart count 2 Nov 26 05:27:52.726: INFO: hostexec-bootstrap-e2e-minion-group-vhdj-s5g5n started at 2022-11-26 05:23:20 +0000 UTC (0+1 container statuses recorded) Nov 26 05:27:52.726: INFO: Container agnhost-container ready: false, restart count 2 Nov 26 05:27:52.726: INFO: mutability-test-7ktz4 started at 2022-11-26 05:23:40 +0000 UTC (0+1 container statuses recorded) Nov 26 05:27:52.726: INFO: Container netexec ready: true, restart count 2 Nov 26 05:27:52.726: INFO: konnectivity-agent-nt4fd started at 2022-11-26 05:05:39 +0000 UTC (0+1 container statuses recorded) Nov 26 05:27:52.726: INFO: Container konnectivity-agent ready: true, restart count 7 Nov 26 05:27:52.726: INFO: csi-hostpathplugin-0 started at 2022-11-26 05:10:50 +0000 UTC (0+7 container statuses recorded) Nov 26 05:27:52.726: INFO: Container csi-attacher ready: true, restart count 6 Nov 26 05:27:52.726: INFO: Container csi-provisioner ready: true, restart count 6 Nov 26 05:27:52.726: INFO: Container csi-resizer ready: true, restart count 6 Nov 26 05:27:52.726: INFO: Container csi-snapshotter ready: true, restart count 6 Nov 26 05:27:52.726: INFO: Container hostpath ready: true, restart count 6 Nov 26 05:27:52.726: INFO: Container liveness-probe ready: true, restart count 6 Nov 26 05:27:52.726: INFO: Container node-driver-registrar ready: true, restart count 6 Nov 26 05:27:52.726: INFO: test-container-pod started at 2022-11-26 05:12:10 +0000 UTC (0+1 container statuses recorded) Nov 26 05:27:52.726: INFO: Container webserver ready: true, restart count 4 Nov 26 05:27:52.726: INFO: csi-hostpathplugin-0 started at 2022-11-26 05:13:07 +0000 UTC (0+7 container statuses recorded) Nov 26 05:27:52.726: INFO: Container csi-attacher ready: true, restart count 3 Nov 26 05:27:52.726: INFO: Container csi-provisioner ready: true, restart count 3 Nov 26 05:27:52.726: INFO: Container csi-resizer ready: true, restart count 3 Nov 26 05:27:52.726: INFO: Container csi-snapshotter ready: true, restart count 3 Nov 26 05:27:52.726: INFO: Container hostpath ready: true, restart count 3 Nov 26 05:27:52.726: INFO: Container liveness-probe ready: true, restart count 3 Nov 26 05:27:52.726: INFO: Container node-driver-registrar ready: true, restart count 3 Nov 26 05:27:52.726: INFO: hostexec-bootstrap-e2e-minion-group-vhdj-p7958 started at 2022-11-26 05:22:59 +0000 UTC (0+1 container statuses recorded) Nov 26 05:27:52.726: INFO: Container agnhost-container ready: true, restart count 3 Nov 26 05:27:52.726: INFO: affinity-lb-esipp-cxg4f started at 2022-11-26 05:23:40 +0000 UTC (0+1 container statuses recorded) Nov 26 05:27:52.726: INFO: Container affinity-lb-esipp ready: true, restart count 0 Nov 26 05:27:52.726: INFO: test-hostpath-type-9rfsr started at 2022-11-26 05:23:17 +0000 UTC (0+1 container statuses recorded) Nov 26 05:27:52.726: INFO: Container host-path-testing ready: false, restart count 0 Nov 26 05:27:52.726: INFO: pvc-tester-pcd6m started at 2022-11-26 05:22:59 +0000 UTC (0+1 container statuses recorded) Nov 26 05:27:52.726: INFO: Container write-pod ready: false, restart count 0 Nov 26 05:27:52.726: INFO: failure-2 started at 2022-11-26 05:16:32 +0000 UTC (0+1 container statuses recorded) Nov 26 05:27:52.726: INFO: Container failure-2 ready: true, restart count 3 Nov 26 05:27:52.726: INFO: external-local-pods-4s55w started at 2022-11-26 05:23:41 +0000 UTC (0+1 container statuses recorded) Nov 26 05:27:52.726: INFO: Container netexec ready: true, restart count 0 Nov 26 05:27:52.726: INFO: mutability-test-55dcq started at 2022-11-26 05:13:49 +0000 UTC (0+1 container statuses recorded) Nov 26 05:27:52.726: INFO: Container netexec ready: true, restart count 5 Nov 26 05:27:52.726: INFO: test-hostpath-type-4c757 started at 2022-11-26 05:23:39 +0000 UTC (0+1 container statuses recorded) Nov 26 05:27:52.726: INFO: Container host-path-testing ready: true, restart count 0 Nov 26 05:27:52.726: INFO: netserver-2 started at 2022-11-26 05:09:51 +0000 UTC (0+1 container statuses recorded) Nov 26 05:27:52.726: INFO: Container webserver ready: true, restart count 5 Nov 26 05:27:52.726: INFO: test-hostpath-type-8kd95 started at 2022-11-26 05:23:45 +0000 UTC (0+1 container statuses recorded) Nov 26 05:27:52.726: INFO: Container host-path-testing ready: false, restart count 0 Nov 26 05:27:52.726: INFO: netserver-2 started at 2022-11-26 05:11:06 +0000 UTC (0+1 container statuses recorded) Nov 26 05:27:52.726: INFO: Container webserver ready: true, restart count 2 Nov 26 05:27:52.726: INFO: netserver-2 started at 2022-11-26 05:11:29 +0000 UTC (0+1 container statuses recorded) Nov 26 05:27:52.726: INFO: Container webserver ready: true, restart count 3 Nov 26 05:27:52.999: INFO: Latency metrics for node bootstrap-e2e-minion-group-vhdj [DeferCleanup (Each)] [sig-network] LoadBalancers ESIPP [Slow] tear down framework | framework.go:193 STEP: Destroying namespace "esipp-9139" for this suite. 11/26/22 05:27:53
Filter through log files | View test history on testgrid
go run hack/e2e.go -v --test --test_args='--ginkgo.focus=Kubernetes\se2e\ssuite\s\[It\]\s\[sig\-network\]\sLoadBalancers\sshould\sbe\sable\sto\schange\sthe\stype\sand\sports\sof\sa\sTCP\sservice\s\[Slow\]$'
test/e2e/network/loadbalancer.go:272 k8s.io/kubernetes/test/e2e/network.glob..func19.3() test/e2e/network/loadbalancer.go:272 +0x17c5 There were additional failures detected after the initial failure: [FAILED] Nov 26 05:13:57.920: failed to list events in namespace "loadbalancers-1625": Get "https://35.227.179.144/api/v1/namespaces/loadbalancers-1625/events": dial tcp 35.227.179.144:443: connect: connection refused In [DeferCleanup (Each)] at: test/e2e/framework/debug/dump.go:44 ---------- [FAILED] Nov 26 05:13:57.960: Couldn't delete ns: "loadbalancers-1625": Delete "https://35.227.179.144/api/v1/namespaces/loadbalancers-1625": dial tcp 35.227.179.144:443: connect: connection refused (&url.Error{Op:"Delete", URL:"https://35.227.179.144/api/v1/namespaces/loadbalancers-1625", Err:(*net.OpError)(0xc003924320)}) In [DeferCleanup (Each)] at: test/e2e/framework/framework.go:370from junit_01.xml
[BeforeEach] [sig-network] LoadBalancers set up framework | framework.go:178 STEP: Creating a kubernetes client 11/26/22 05:07:54.066 Nov 26 05:07:54.066: INFO: >>> kubeConfig: /workspace/.kube/config STEP: Building a namespace api object, basename loadbalancers 11/26/22 05:07:54.067 STEP: Waiting for a default service account to be provisioned in namespace 11/26/22 05:07:54.19 STEP: Waiting for kube-root-ca.crt to be provisioned in namespace 11/26/22 05:07:54.276 [BeforeEach] [sig-network] LoadBalancers test/e2e/framework/metrics/init/init.go:31 [BeforeEach] [sig-network] LoadBalancers test/e2e/network/loadbalancer.go:65 [It] should be able to change the type and ports of a TCP service [Slow] test/e2e/network/loadbalancer.go:77 Nov 26 05:07:54.448: INFO: namespace for TCP test: loadbalancers-1625 STEP: creating a TCP service mutability-test with type=ClusterIP in namespace loadbalancers-1625 11/26/22 05:07:54.492 Nov 26 05:07:54.540: INFO: service port TCP: 80 STEP: creating a pod to be part of the TCP service mutability-test 11/26/22 05:07:54.54 Nov 26 05:07:54.627: INFO: Waiting up to 2m0s for 1 pods to be created Nov 26 05:07:54.668: INFO: Found all 1 pods Nov 26 05:07:54.668: INFO: Waiting up to 2m0s for 1 pods to be running and ready: [mutability-test-5jd94] Nov 26 05:07:54.668: INFO: Waiting up to 2m0s for pod "mutability-test-5jd94" in namespace "loadbalancers-1625" to be "running and ready" Nov 26 05:07:54.711: INFO: Pod "mutability-test-5jd94": Phase="Pending", Reason="", readiness=false. Elapsed: 42.255182ms Nov 26 05:07:54.711: INFO: Error evaluating pod condition running and ready: want pod 'mutability-test-5jd94' on 'bootstrap-e2e-minion-group-vhdj' to be 'Running' but was 'Pending' Nov 26 05:07:56.754: INFO: Pod "mutability-test-5jd94": Phase="Pending", Reason="", readiness=false. Elapsed: 2.085893003s Nov 26 05:07:56.754: INFO: Error evaluating pod condition running and ready: want pod 'mutability-test-5jd94' on 'bootstrap-e2e-minion-group-vhdj' to be 'Running' but was 'Pending' Nov 26 05:07:58.755: INFO: Pod "mutability-test-5jd94": Phase="Pending", Reason="", readiness=false. Elapsed: 4.086865493s Nov 26 05:07:58.755: INFO: Error evaluating pod condition running and ready: want pod 'mutability-test-5jd94' on 'bootstrap-e2e-minion-group-vhdj' to be 'Running' but was 'Pending' Nov 26 05:08:00.752: INFO: Pod "mutability-test-5jd94": Phase="Pending", Reason="", readiness=false. Elapsed: 6.083450653s Nov 26 05:08:00.752: INFO: Error evaluating pod condition running and ready: want pod 'mutability-test-5jd94' on 'bootstrap-e2e-minion-group-vhdj' to be 'Running' but was 'Pending' Nov 26 05:08:02.752: INFO: Pod "mutability-test-5jd94": Phase="Pending", Reason="", readiness=false. Elapsed: 8.083595735s Nov 26 05:08:02.752: INFO: Error evaluating pod condition running and ready: want pod 'mutability-test-5jd94' on 'bootstrap-e2e-minion-group-vhdj' to be 'Running' but was 'Pending' Nov 26 05:08:04.753: INFO: Pod "mutability-test-5jd94": Phase="Pending", Reason="", readiness=false. Elapsed: 10.084556095s Nov 26 05:08:04.753: INFO: Error evaluating pod condition running and ready: want pod 'mutability-test-5jd94' on 'bootstrap-e2e-minion-group-vhdj' to be 'Running' but was 'Pending' Nov 26 05:08:06.753: INFO: Pod "mutability-test-5jd94": Phase="Pending", Reason="", readiness=false. Elapsed: 12.085038121s Nov 26 05:08:06.753: INFO: Error evaluating pod condition running and ready: want pod 'mutability-test-5jd94' on 'bootstrap-e2e-minion-group-vhdj' to be 'Running' but was 'Pending' Nov 26 05:08:08.752: INFO: Pod "mutability-test-5jd94": Phase="Running", Reason="", readiness=true. Elapsed: 14.084096417s Nov 26 05:08:08.752: INFO: Pod "mutability-test-5jd94" satisfied condition "running and ready" Nov 26 05:08:08.752: INFO: Wanted all 1 pods to be running and ready. Result: true. Pods: [mutability-test-5jd94] STEP: changing the TCP service to type=NodePort 11/26/22 05:08:08.752 Nov 26 05:08:08.840: INFO: TCP node port: 32123 STEP: hitting the TCP service's NodePort 11/26/22 05:08:08.84 Nov 26 05:08:08.840: INFO: Poking "http://35.247.104.236:32123/echo?msg=hello" Nov 26 05:08:08.880: INFO: Poke("http://35.247.104.236:32123/echo?msg=hello"): Get "http://35.247.104.236:32123/echo?msg=hello": dial tcp 35.247.104.236:32123: connect: connection refused Nov 26 05:08:10.881: INFO: Poking "http://35.247.104.236:32123/echo?msg=hello" Nov 26 05:08:10.920: INFO: Poke("http://35.247.104.236:32123/echo?msg=hello"): Get "http://35.247.104.236:32123/echo?msg=hello": dial tcp 35.247.104.236:32123: connect: connection refused Nov 26 05:08:12.881: INFO: Poking "http://35.247.104.236:32123/echo?msg=hello" Nov 26 05:08:12.920: INFO: Poke("http://35.247.104.236:32123/echo?msg=hello"): Get "http://35.247.104.236:32123/echo?msg=hello": dial tcp 35.247.104.236:32123: connect: connection refused Nov 26 05:08:14.881: INFO: Poking "http://35.247.104.236:32123/echo?msg=hello" Nov 26 05:08:14.920: INFO: Poke("http://35.247.104.236:32123/echo?msg=hello"): Get "http://35.247.104.236:32123/echo?msg=hello": dial tcp 35.247.104.236:32123: connect: connection refused Nov 26 05:08:16.880: INFO: Poking "http://35.247.104.236:32123/echo?msg=hello" Nov 26 05:08:16.920: INFO: Poke("http://35.247.104.236:32123/echo?msg=hello"): Get "http://35.247.104.236:32123/echo?msg=hello": dial tcp 35.247.104.236:32123: connect: connection refused Nov 26 05:08:18.881: INFO: Poking "http://35.247.104.236:32123/echo?msg=hello" Nov 26 05:08:18.961: INFO: Poke("http://35.247.104.236:32123/echo?msg=hello"): success STEP: creating a static load balancer IP 11/26/22 05:08:18.961 Nov 26 05:08:20.984: INFO: Allocated static load balancer IP: 35.185.230.152 STEP: changing the TCP service to type=LoadBalancer 11/26/22 05:08:20.984 STEP: waiting for the TCP service to have a load balancer 11/26/22 05:08:21.107 Nov 26 05:08:21.107: INFO: Waiting up to 15m0s for service "mutability-test" to have a LoadBalancer Nov 26 05:10:09.192: INFO: TCP load balancer: 35.185.230.152 STEP: demoting the static IP to ephemeral 11/26/22 05:10:09.192 STEP: hitting the TCP service's NodePort 11/26/22 05:10:10.823 Nov 26 05:10:10.823: INFO: Poking "http://35.247.104.236:32123/echo?msg=hello" Nov 26 05:10:10.904: INFO: Poke("http://35.247.104.236:32123/echo?msg=hello"): success STEP: hitting the TCP service's LoadBalancer 11/26/22 05:10:10.904 Nov 26 05:10:10.904: INFO: Poking "http://35.185.230.152:80/echo?msg=hello" Nov 26 05:10:18.191: INFO: Poke("http://35.185.230.152:80/echo?msg=hello"): success STEP: changing the TCP service's NodePort 11/26/22 05:10:18.191 Nov 26 05:10:18.287: INFO: TCP node port: 32124 STEP: hitting the TCP service's new NodePort 11/26/22 05:10:18.287 Nov 26 05:10:18.287: INFO: Poking "http://35.247.104.236:32124/echo?msg=hello" Nov 26 05:10:18.327: INFO: Poke("http://35.247.104.236:32124/echo?msg=hello"): Get "http://35.247.104.236:32124/echo?msg=hello": dial tcp 35.247.104.236:32124: connect: connection refused Nov 26 05:10:20.329: INFO: Poking "http://35.247.104.236:32124/echo?msg=hello" Nov 26 05:10:20.368: INFO: Poke("http://35.247.104.236:32124/echo?msg=hello"): Get "http://35.247.104.236:32124/echo?msg=hello": dial tcp 35.247.104.236:32124: connect: connection refused Nov 26 05:10:22.328: INFO: Poking "http://35.247.104.236:32124/echo?msg=hello" Nov 26 05:10:22.367: INFO: Poke("http://35.247.104.236:32124/echo?msg=hello"): Get "http://35.247.104.236:32124/echo?msg=hello": dial tcp 35.247.104.236:32124: connect: connection refused Nov 26 05:10:24.328: INFO: Poking "http://35.247.104.236:32124/echo?msg=hello" Nov 26 05:10:24.367: INFO: Poke("http://35.247.104.236:32124/echo?msg=hello"): Get "http://35.247.104.236:32124/echo?msg=hello": dial tcp 35.247.104.236:32124: connect: connection refused Nov 26 05:10:26.328: INFO: Poking "http://35.247.104.236:32124/echo?msg=hello" Nov 26 05:10:26.408: INFO: Poke("http://35.247.104.236:32124/echo?msg=hello"): success STEP: checking the old TCP NodePort is closed 11/26/22 05:10:26.408 Nov 26 05:10:26.408: INFO: Poking "http://35.247.104.236:32123/" Nov 26 05:10:26.447: INFO: Poke("http://35.247.104.236:32123/"): Get "http://35.247.104.236:32123/": dial tcp 35.247.104.236:32123: connect: connection refused STEP: hitting the TCP service's LoadBalancer 11/26/22 05:10:26.447 Nov 26 05:10:26.447: INFO: Poking "http://35.185.230.152:80/echo?msg=hello" Nov 26 05:10:26.528: INFO: Poke("http://35.185.230.152:80/echo?msg=hello"): success STEP: changing the TCP service's port 11/26/22 05:10:26.528 Nov 26 05:10:26.638: INFO: service port TCP: 81 STEP: hitting the TCP service's NodePort 11/26/22 05:10:26.638 Nov 26 05:10:26.638: INFO: Poking "http://35.247.104.236:32124/echo?msg=hello" Nov 26 05:10:26.726: INFO: Poke("http://35.247.104.236:32124/echo?msg=hello"): success STEP: hitting the TCP service's LoadBalancer 11/26/22 05:10:26.726 Nov 26 05:10:26.726: INFO: Poking "http://35.185.230.152:81/echo?msg=hello" Nov 26 05:10:36.727: INFO: Poke("http://35.185.230.152:81/echo?msg=hello"): Get "http://35.185.230.152:81/echo?msg=hello": context deadline exceeded (Client.Timeout exceeded while awaiting headers) Nov 26 05:10:38.728: INFO: Poking "http://35.185.230.152:81/echo?msg=hello" Nov 26 05:10:48.728: INFO: Poke("http://35.185.230.152:81/echo?msg=hello"): Get "http://35.185.230.152:81/echo?msg=hello": context deadline exceeded (Client.Timeout exceeded while awaiting headers) Nov 26 05:10:50.728: INFO: Poking "http://35.185.230.152:81/echo?msg=hello" Nov 26 05:11:00.728: INFO: Poke("http://35.185.230.152:81/echo?msg=hello"): Get "http://35.185.230.152:81/echo?msg=hello": context deadline exceeded (Client.Timeout exceeded while awaiting headers) Nov 26 05:11:02.727: INFO: Poking "http://35.185.230.152:81/echo?msg=hello" Nov 26 05:11:12.728: INFO: Poke("http://35.185.230.152:81/echo?msg=hello"): Get "http://35.185.230.152:81/echo?msg=hello": context deadline exceeded (Client.Timeout exceeded while awaiting headers) Nov 26 05:11:14.727: INFO: Poking "http://35.185.230.152:81/echo?msg=hello" Nov 26 05:11:24.727: INFO: Poke("http://35.185.230.152:81/echo?msg=hello"): Get "http://35.185.230.152:81/echo?msg=hello": context deadline exceeded (Client.Timeout exceeded while awaiting headers) Nov 26 05:11:26.727: INFO: Poking "http://35.185.230.152:81/echo?msg=hello" Nov 26 05:11:36.728: INFO: Poke("http://35.185.230.152:81/echo?msg=hello"): Get "http://35.185.230.152:81/echo?msg=hello": context deadline exceeded (Client.Timeout exceeded while awaiting headers) Nov 26 05:11:36.728: INFO: Poking "http://35.185.230.152:81/echo?msg=hello" Nov 26 05:11:46.729: INFO: Poke("http://35.185.230.152:81/echo?msg=hello"): Get "http://35.185.230.152:81/echo?msg=hello": context deadline exceeded (Client.Timeout exceeded while awaiting headers) Nov 26 05:11:48.728: INFO: Poking "http://35.185.230.152:81/echo?msg=hello" Nov 26 05:11:58.729: INFO: Poke("http://35.185.230.152:81/echo?msg=hello"): Get "http://35.185.230.152:81/echo?msg=hello": context deadline exceeded (Client.Timeout exceeded while awaiting headers) Nov 26 05:12:00.727: INFO: Poking "http://35.185.230.152:81/echo?msg=hello" Nov 26 05:12:10.727: INFO: Poke("http://35.185.230.152:81/echo?msg=hello"): Get "http://35.185.230.152:81/echo?msg=hello": context deadline exceeded (Client.Timeout exceeded while awaiting headers) Nov 26 05:12:12.728: INFO: Poking "http://35.185.230.152:81/echo?msg=hello" Nov 26 05:12:22.728: INFO: Poke("http://35.185.230.152:81/echo?msg=hello"): Get "http://35.185.230.152:81/echo?msg=hello": context deadline exceeded (Client.Timeout exceeded while awaiting headers) Nov 26 05:12:24.728: INFO: Poking "http://35.185.230.152:81/echo?msg=hello" Nov 26 05:12:34.728: INFO: Poke("http://35.185.230.152:81/echo?msg=hello"): Get "http://35.185.230.152:81/echo?msg=hello": context deadline exceeded (Client.Timeout exceeded while awaiting headers) Nov 26 05:12:36.728: INFO: Poking "http://35.185.230.152:81/echo?msg=hello" Nov 26 05:12:36.769: INFO: Poke("http://35.185.230.152:81/echo?msg=hello"): Get "http://35.185.230.152:81/echo?msg=hello": dial tcp 35.185.230.152:81: connect: connection refused Nov 26 05:12:38.727: INFO: Poking "http://35.185.230.152:81/echo?msg=hello" Nov 26 05:12:38.766: INFO: Poke("http://35.185.230.152:81/echo?msg=hello"): Get "http://35.185.230.152:81/echo?msg=hello": dial tcp 35.185.230.152:81: connect: connection refused Nov 26 05:12:40.727: INFO: Poking "http://35.185.230.152:81/echo?msg=hello" Nov 26 05:12:40.767: INFO: Poke("http://35.185.230.152:81/echo?msg=hello"): Get "http://35.185.230.152:81/echo?msg=hello": dial tcp 35.185.230.152:81: connect: connection refused Nov 26 05:12:42.727: INFO: Poking "http://35.185.230.152:81/echo?msg=hello" Nov 26 05:12:42.767: INFO: Poke("http://35.185.230.152:81/echo?msg=hello"): Get "http://35.185.230.152:81/echo?msg=hello": dial tcp 35.185.230.152:81: connect: connection refused Nov 26 05:12:44.727: INFO: Poking "http://35.185.230.152:81/echo?msg=hello" Nov 26 05:12:44.767: INFO: Poke("http://35.185.230.152:81/echo?msg=hello"): Get "http://35.185.230.152:81/echo?msg=hello": dial tcp 35.185.230.152:81: connect: connection refused Nov 26 05:12:46.727: INFO: Poking "http://35.185.230.152:81/echo?msg=hello" Nov 26 05:12:46.767: INFO: Poke("http://35.185.230.152:81/echo?msg=hello"): Get "http://35.185.230.152:81/echo?msg=hello": dial tcp 35.185.230.152:81: connect: connection refused Nov 26 05:12:48.727: INFO: Poking "http://35.185.230.152:81/echo?msg=hello" Nov 26 05:12:48.766: INFO: Poke("http://35.185.230.152:81/echo?msg=hello"): Get "http://35.185.230.152:81/echo?msg=hello": dial tcp 35.185.230.152:81: connect: connection refused Nov 26 05:12:50.728: INFO: Poking "http://35.185.230.152:81/echo?msg=hello" Nov 26 05:12:50.767: INFO: Poke("http://35.185.230.152:81/echo?msg=hello"): Get "http://35.185.230.152:81/echo?msg=hello": dial tcp 35.185.230.152:81: connect: connection refused Nov 26 05:12:52.728: INFO: Poking "http://35.185.230.152:81/echo?msg=hello" Nov 26 05:12:52.767: INFO: Poke("http://35.185.230.152:81/echo?msg=hello"): Get "http://35.185.230.152:81/echo?msg=hello": dial tcp 35.185.230.152:81: connect: connection refused ------------------------------ Progress Report for Ginkgo Process #20 Automatically polling progress: [sig-network] LoadBalancers should be able to change the type and ports of a TCP service [Slow] (Spec Runtime: 5m0.338s) test/e2e/network/loadbalancer.go:77 In [It] (Node Runtime: 5m0.001s) test/e2e/network/loadbalancer.go:77 At [By Step] hitting the TCP service's LoadBalancer (Step Runtime: 2m27.677s) test/e2e/network/loadbalancer.go:243 Spec Goroutine goroutine 600 [select] k8s.io/kubernetes/vendor/k8s.io/apimachinery/pkg/util/wait.WaitForWithContext({0x7fe0bc8, 0xc0000820c8}, 0xc002914ea0, 0x2fdb16a?) vendor/k8s.io/apimachinery/pkg/util/wait/wait.go:660 k8s.io/kubernetes/vendor/k8s.io/apimachinery/pkg/util/wait.poll({0x7fe0bc8, 0xc0000820c8}, 0xd0?, 0x2fd9d05?, 0x28?) vendor/k8s.io/apimachinery/pkg/util/wait/wait.go:596 k8s.io/kubernetes/vendor/k8s.io/apimachinery/pkg/util/wait.PollImmediateWithContext({0x7fe0bc8, 0xc0000820c8}, 0x30?, 0xc0015bdc20?, 0x262a967?) vendor/k8s.io/apimachinery/pkg/util/wait/wait.go:528 k8s.io/kubernetes/vendor/k8s.io/apimachinery/pkg/util/wait.PollImmediate(0x754e980?, 0xc000def8a8?, 0x76888de?) vendor/k8s.io/apimachinery/pkg/util/wait/wait.go:514 k8s.io/kubernetes/test/e2e/framework/service.TestReachableHTTPWithRetriableErrorCodes({0xc002559570, 0xe}, 0x51, {0xae73300, 0x0, 0x0}, 0x1?) test/e2e/framework/service/util.go:46 k8s.io/kubernetes/test/e2e/framework/service.TestReachableHTTP(...) test/e2e/framework/service/util.go:29 > k8s.io/kubernetes/test/e2e/network.glob..func19.3() test/e2e/network/loadbalancer.go:244 k8s.io/kubernetes/vendor/github.com/onsi/ginkgo/v2/internal.extractBodyFunction.func3({0x2d591ce, 0xc001dea780}) vendor/github.com/onsi/ginkgo/v2/internal/node.go:449 k8s.io/kubernetes/vendor/github.com/onsi/ginkgo/v2/internal.(*Suite).runNode.func2() vendor/github.com/onsi/ginkgo/v2/internal/suite.go:750 k8s.io/kubernetes/vendor/github.com/onsi/ginkgo/v2/internal.(*Suite).runNode vendor/github.com/onsi/ginkgo/v2/internal/suite.go:738 ------------------------------ Nov 26 05:12:54.728: INFO: Poking "http://35.185.230.152:81/echo?msg=hello" Nov 26 05:12:54.767: INFO: Poke("http://35.185.230.152:81/echo?msg=hello"): Get "http://35.185.230.152:81/echo?msg=hello": dial tcp 35.185.230.152:81: connect: connection refused Nov 26 05:12:56.727: INFO: Poking "http://35.185.230.152:81/echo?msg=hello" Nov 26 05:12:56.766: INFO: Poke("http://35.185.230.152:81/echo?msg=hello"): Get "http://35.185.230.152:81/echo?msg=hello": dial tcp 35.185.230.152:81: connect: connection refused Nov 26 05:12:58.727: INFO: Poking "http://35.185.230.152:81/echo?msg=hello" Nov 26 05:12:58.767: INFO: Poke("http://35.185.230.152:81/echo?msg=hello"): Get "http://35.185.230.152:81/echo?msg=hello": dial tcp 35.185.230.152:81: connect: connection refused Nov 26 05:13:00.727: INFO: Poking "http://35.185.230.152:81/echo?msg=hello" Nov 26 05:13:00.766: INFO: Poke("http://35.185.230.152:81/echo?msg=hello"): Get "http://35.185.230.152:81/echo?msg=hello": dial tcp 35.185.230.152:81: connect: connection refused Nov 26 05:13:02.727: INFO: Poking "http://35.185.230.152:81/echo?msg=hello" Nov 26 05:13:02.767: INFO: Poke("http://35.185.230.152:81/echo?msg=hello"): Get "http://35.185.230.152:81/echo?msg=hello": dial tcp 35.185.230.152:81: connect: connection refused Nov 26 05:13:04.727: INFO: Poking "http://35.185.230.152:81/echo?msg=hello" Nov 26 05:13:04.766: INFO: Poke("http://35.185.230.152:81/echo?msg=hello"): Get "http://35.185.230.152:81/echo?msg=hello": dial tcp 35.185.230.152:81: connect: connection refused Nov 26 05:13:06.728: INFO: Poking "http://35.185.230.152:81/echo?msg=hello" Nov 26 05:13:06.768: INFO: Poke("http://35.185.230.152:81/echo?msg=hello"): Get "http://35.185.230.152:81/echo?msg=hello": dial tcp 35.185.230.152:81: connect: connection refused Nov 26 05:13:08.727: INFO: Poking "http://35.185.230.152:81/echo?msg=hello" Nov 26 05:13:08.767: INFO: Poke("http://35.185.230.152:81/echo?msg=hello"): Get "http://35.185.230.152:81/echo?msg=hello": dial tcp 35.185.230.152:81: connect: connection refused Nov 26 05:13:10.727: INFO: Poking "http://35.185.230.152:81/echo?msg=hello" Nov 26 05:13:10.766: INFO: Poke("http://35.185.230.152:81/echo?msg=hello"): Get "http://35.185.230.152:81/echo?msg=hello": dial tcp 35.185.230.152:81: connect: connection refused Nov 26 05:13:12.727: INFO: Poking "http://35.185.230.152:81/echo?msg=hello" Nov 26 05:13:12.767: INFO: Poke("http://35.185.230.152:81/echo?msg=hello"): Get "http://35.185.230.152:81/echo?msg=hello": dial tcp 35.185.230.152:81: connect: connection refused ------------------------------ Progress Report for Ginkgo Process #20 Automatically polling progress: [sig-network] LoadBalancers should be able to change the type and ports of a TCP service [Slow] (Spec Runtime: 5m20.34s) test/e2e/network/loadbalancer.go:77 In [It] (Node Runtime: 5m20.003s) test/e2e/network/loadbalancer.go:77 At [By Step] hitting the TCP service's LoadBalancer (Step Runtime: 2m47.679s) test/e2e/network/loadbalancer.go:243 Spec Goroutine goroutine 600 [select] k8s.io/kubernetes/vendor/k8s.io/apimachinery/pkg/util/wait.WaitForWithContext({0x7fe0bc8, 0xc0000820c8}, 0xc002914ea0, 0x2fdb16a?) vendor/k8s.io/apimachinery/pkg/util/wait/wait.go:660 k8s.io/kubernetes/vendor/k8s.io/apimachinery/pkg/util/wait.poll({0x7fe0bc8, 0xc0000820c8}, 0xd0?, 0x2fd9d05?, 0x28?) vendor/k8s.io/apimachinery/pkg/util/wait/wait.go:596 k8s.io/kubernetes/vendor/k8s.io/apimachinery/pkg/util/wait.PollImmediateWithContext({0x7fe0bc8, 0xc0000820c8}, 0x30?, 0xc0015bdc20?, 0x262a967?) vendor/k8s.io/apimachinery/pkg/util/wait/wait.go:528 k8s.io/kubernetes/vendor/k8s.io/apimachinery/pkg/util/wait.PollImmediate(0x754e980?, 0xc000def8a8?, 0x76888de?) vendor/k8s.io/apimachinery/pkg/util/wait/wait.go:514 k8s.io/kubernetes/test/e2e/framework/service.TestReachableHTTPWithRetriableErrorCodes({0xc002559570, 0xe}, 0x51, {0xae73300, 0x0, 0x0}, 0x1?) test/e2e/framework/service/util.go:46 k8s.io/kubernetes/test/e2e/framework/service.TestReachableHTTP(...) test/e2e/framework/service/util.go:29 > k8s.io/kubernetes/test/e2e/network.glob..func19.3() test/e2e/network/loadbalancer.go:244 k8s.io/kubernetes/vendor/github.com/onsi/ginkgo/v2/internal.extractBodyFunction.func3({0x2d591ce, 0xc001dea780}) vendor/github.com/onsi/ginkgo/v2/internal/node.go:449 k8s.io/kubernetes/vendor/github.com/onsi/ginkgo/v2/internal.(*Suite).runNode.func2() vendor/github.com/onsi/ginkgo/v2/internal/suite.go:750 k8s.io/kubernetes/vendor/github.com/onsi/ginkgo/v2/internal.(*Suite).runNode vendor/github.com/onsi/ginkgo/v2/internal/suite.go:738 ------------------------------ Nov 26 05:13:14.728: INFO: Poking "http://35.185.230.152:81/echo?msg=hello" Nov 26 05:13:14.767: INFO: Poke("http://35.185.230.152:81/echo?msg=hello"): Get "http://35.185.230.152:81/echo?msg=hello": dial tcp 35.185.230.152:81: connect: connection refused Nov 26 05:13:16.727: INFO: Poking "http://35.185.230.152:81/echo?msg=hello" Nov 26 05:13:16.767: INFO: Poke("http://35.185.230.152:81/echo?msg=hello"): Get "http://35.185.230.152:81/echo?msg=hello": dial tcp 35.185.230.152:81: connect: connection refused Nov 26 05:13:18.728: INFO: Poking "http://35.185.230.152:81/echo?msg=hello" Nov 26 05:13:18.768: INFO: Poke("http://35.185.230.152:81/echo?msg=hello"): Get "http://35.185.230.152:81/echo?msg=hello": dial tcp 35.185.230.152:81: connect: connection refused Nov 26 05:13:20.728: INFO: Poking "http://35.185.230.152:81/echo?msg=hello" Nov 26 05:13:20.767: INFO: Poke("http://35.185.230.152:81/echo?msg=hello"): Get "http://35.185.230.152:81/echo?msg=hello": dial tcp 35.185.230.152:81: connect: connection refused Nov 26 05:13:22.728: INFO: Poking "http://35.185.230.152:81/echo?msg=hello" Nov 26 05:13:22.767: INFO: Poke("http://35.185.230.152:81/echo?msg=hello"): Get "http://35.185.230.152:81/echo?msg=hello": dial tcp 35.185.230.152:81: connect: connection refused Nov 26 05:13:24.727: INFO: Poking "http://35.185.230.152:81/echo?msg=hello" Nov 26 05:13:24.766: INFO: Poke("http://35.185.230.152:81/echo?msg=hello"): Get "http://35.185.230.152:81/echo?msg=hello": dial tcp 35.185.230.152:81: connect: connection refused Nov 26 05:13:26.727: INFO: Poking "http://35.185.230.152:81/echo?msg=hello" Nov 26 05:13:26.766: INFO: Poke("http://35.185.230.152:81/echo?msg=hello"): Get "http://35.185.230.152:81/echo?msg=hello": dial tcp 35.185.230.152:81: connect: connection refused Nov 26 05:13:28.727: INFO: Poking "http://35.185.230.152:81/echo?msg=hello" Nov 26 05:13:28.766: INFO: Poke("http://35.185.230.152:81/echo?msg=hello"): Get "http://35.185.230.152:81/echo?msg=hello": dial tcp 35.185.230.152:81: connect: connection refused Nov 26 05:13:30.727: INFO: Poking "http://35.185.230.152:81/echo?msg=hello" Nov 26 05:13:30.768: INFO: Poke("http://35.185.230.152:81/echo?msg=hello"): Get "http://35.185.230.152:81/echo?msg=hello": dial tcp 35.185.230.152:81: connect: connection refused Nov 26 05:13:32.728: INFO: Poking "http://35.185.230.152:81/echo?msg=hello" Nov 26 05:13:32.767: INFO: Poke("http://35.185.230.152:81/echo?msg=hello"): Get "http://35.185.230.152:81/echo?msg=hello": dial tcp 35.185.230.152:81: connect: connection refused ------------------------------ Progress Report for Ginkgo Process #20 Automatically polling progress: [sig-network] LoadBalancers should be able to change the type and ports of a TCP service [Slow] (Spec Runtime: 5m40.342s) test/e2e/network/loadbalancer.go:77 In [It] (Node Runtime: 5m40.005s) test/e2e/network/loadbalancer.go:77 At [By Step] hitting the TCP service's LoadBalancer (Step Runtime: 3m7.682s) test/e2e/network/loadbalancer.go:243 Spec Goroutine goroutine 600 [select] k8s.io/kubernetes/vendor/k8s.io/apimachinery/pkg/util/wait.WaitForWithContext({0x7fe0bc8, 0xc0000820c8}, 0xc002914ea0, 0x2fdb16a?) vendor/k8s.io/apimachinery/pkg/util/wait/wait.go:660 k8s.io/kubernetes/vendor/k8s.io/apimachinery/pkg/util/wait.poll({0x7fe0bc8, 0xc0000820c8}, 0xd0?, 0x2fd9d05?, 0x28?) vendor/k8s.io/apimachinery/pkg/util/wait/wait.go:596 k8s.io/kubernetes/vendor/k8s.io/apimachinery/pkg/util/wait.PollImmediateWithContext({0x7fe0bc8, 0xc0000820c8}, 0x30?, 0xc0015bdc20?, 0x262a967?) vendor/k8s.io/apimachinery/pkg/util/wait/wait.go:528 k8s.io/kubernetes/vendor/k8s.io/apimachinery/pkg/util/wait.PollImmediate(0x754e980?, 0xc000def8a8?, 0x76888de?) vendor/k8s.io/apimachinery/pkg/util/wait/wait.go:514 k8s.io/kubernetes/test/e2e/framework/service.TestReachableHTTPWithRetriableErrorCodes({0xc002559570, 0xe}, 0x51, {0xae73300, 0x0, 0x0}, 0x1?) test/e2e/framework/service/util.go:46 k8s.io/kubernetes/test/e2e/framework/service.TestReachableHTTP(...) test/e2e/framework/service/util.go:29 > k8s.io/kubernetes/test/e2e/network.glob..func19.3() test/e2e/network/loadbalancer.go:244 k8s.io/kubernetes/vendor/github.com/onsi/ginkgo/v2/internal.extractBodyFunction.func3({0x2d591ce, 0xc001dea780}) vendor/github.com/onsi/ginkgo/v2/internal/node.go:449 k8s.io/kubernetes/vendor/github.com/onsi/ginkgo/v2/internal.(*Suite).runNode.func2() vendor/github.com/onsi/ginkgo/v2/internal/suite.go:750 k8s.io/kubernetes/vendor/github.com/onsi/ginkgo/v2/internal.(*Suite).runNode vendor/github.com/onsi/ginkgo/v2/internal/suite.go:738 ------------------------------ Nov 26 05:13:34.728: INFO: Poking "http://35.185.230.152:81/echo?msg=hello" Nov 26 05:13:34.767: INFO: Poke("http://35.185.230.152:81/echo?msg=hello"): Get "http://35.185.230.152:81/echo?msg=hello": dial tcp 35.185.230.152:81: connect: connection refused Nov 26 05:13:36.727: INFO: Poking "http://35.185.230.152:81/echo?msg=hello" Nov 26 05:13:36.767: INFO: Poke("http://35.185.230.152:81/echo?msg=hello"): Get "http://35.185.230.152:81/echo?msg=hello": dial tcp 35.185.230.152:81: connect: connection refused Nov 26 05:13:38.727: INFO: Poking "http://35.185.230.152:81/echo?msg=hello" Nov 26 05:13:38.766: INFO: Poke("http://35.185.230.152:81/echo?msg=hello"): Get "http://35.185.230.152:81/echo?msg=hello": dial tcp 35.185.230.152:81: connect: connection refused Nov 26 05:13:40.728: INFO: Poking "http://35.185.230.152:81/echo?msg=hello" Nov 26 05:13:40.768: INFO: Poke("http://35.185.230.152:81/echo?msg=hello"): Get "http://35.185.230.152:81/echo?msg=hello": dial tcp 35.185.230.152:81: connect: connection refused Nov 26 05:13:42.727: INFO: Poking "http://35.185.230.152:81/echo?msg=hello" Nov 26 05:13:42.767: INFO: Poke("http://35.185.230.152:81/echo?msg=hello"): Get "http://35.185.230.152:81/echo?msg=hello": dial tcp 35.185.230.152:81: connect: connection refused Nov 26 05:13:44.727: INFO: Poking "http://35.185.230.152:81/echo?msg=hello" Nov 26 05:13:44.766: INFO: Poke("http://35.185.230.152:81/echo?msg=hello"): Get "http://35.185.230.152:81/echo?msg=hello": dial tcp 35.185.230.152:81: connect: connection refused Nov 26 05:13:46.728: INFO: Poking "http://35.185.230.152:81/echo?msg=hello" Nov 26 05:13:46.767: INFO: Poke("http://35.185.230.152:81/echo?msg=hello"): Get "http://35.185.230.152:81/echo?msg=hello": dial tcp 35.185.230.152:81: connect: connection refused Nov 26 05:13:48.727: INFO: Poking "http://35.185.230.152:81/echo?msg=hello" Nov 26 05:13:48.807: INFO: Poke("http://35.185.230.152:81/echo?msg=hello"): success STEP: Scaling the pods to 0 11/26/22 05:13:48.807 Nov 26 05:13:49.018: INFO: Waiting up to 2m0s for 0 pods to be created Nov 26 05:13:49.083: INFO: Found all 0 pods Nov 26 05:13:49.084: INFO: Waiting up to 2m0s for 0 pods to be running and ready: [] Nov 26 05:13:49.084: INFO: Wanted all 0 pods to be running and ready. Result: true. Pods: [] STEP: looking for ICMP REJECT on the TCP service's NodePort 11/26/22 05:13:49.084 Nov 26 05:13:49.084: INFO: Poking "http://35.247.104.236:32124/" Nov 26 05:13:49.125: INFO: Poke("http://35.247.104.236:32124/"): Get "http://35.247.104.236:32124/": dial tcp 35.247.104.236:32124: connect: connection refused STEP: looking for ICMP REJECT on the TCP service's LoadBalancer 11/26/22 05:13:49.125 Nov 26 05:13:49.125: INFO: Poking "http://35.185.230.152:81/" Nov 26 05:13:49.165: INFO: Poke("http://35.185.230.152:81/"): Get "http://35.185.230.152:81/": dial tcp 35.185.230.152:81: connect: connection refused STEP: Scaling the pods to 1 11/26/22 05:13:49.166 Nov 26 05:13:49.287: INFO: Waiting up to 2m0s for 1 pods to be created Nov 26 05:13:49.347: INFO: Found all 1 pods Nov 26 05:13:49.347: INFO: Waiting up to 2m0s for 1 pods to be running and ready: [mutability-test-55dcq] Nov 26 05:13:49.347: INFO: Waiting up to 2m0s for pod "mutability-test-55dcq" in namespace "loadbalancers-1625" to be "running and ready" Nov 26 05:13:49.428: INFO: Pod "mutability-test-55dcq": Phase="Pending", Reason="", readiness=false. Elapsed: 80.817146ms Nov 26 05:13:49.428: INFO: Error evaluating pod condition running and ready: want pod 'mutability-test-55dcq' on 'bootstrap-e2e-minion-group-vhdj' to be 'Running' but was 'Pending' Nov 26 05:13:51.475: INFO: Pod "mutability-test-55dcq": Phase="Running", Reason="", readiness=true. Elapsed: 2.128471112s Nov 26 05:13:51.475: INFO: Pod "mutability-test-55dcq" satisfied condition "running and ready" Nov 26 05:13:51.475: INFO: Wanted all 1 pods to be running and ready. Result: true. Pods: [mutability-test-55dcq] STEP: hitting the TCP service's NodePort 11/26/22 05:13:51.475 Nov 26 05:13:51.476: INFO: Poking "http://35.247.104.236:32124/echo?msg=hello" Nov 26 05:13:51.518: INFO: Poke("http://35.247.104.236:32124/echo?msg=hello"): Get "http://35.247.104.236:32124/echo?msg=hello": dial tcp 35.247.104.236:32124: connect: no route to host Nov 26 05:13:53.518: INFO: Poking "http://35.247.104.236:32124/echo?msg=hello" Nov 26 05:13:53.559: INFO: Poke("http://35.247.104.236:32124/echo?msg=hello"): Get "http://35.247.104.236:32124/echo?msg=hello": dial tcp 35.247.104.236:32124: connect: no route to host ------------------------------ Progress Report for Ginkgo Process #20 Automatically polling progress: [sig-network] LoadBalancers should be able to change the type and ports of a TCP service [Slow] (Spec Runtime: 6m0.344s) test/e2e/network/loadbalancer.go:77 In [It] (Node Runtime: 6m0.007s) test/e2e/network/loadbalancer.go:77 At [By Step] hitting the TCP service's NodePort (Step Runtime: 2.934s) test/e2e/network/loadbalancer.go:260 Spec Goroutine goroutine 600 [select] k8s.io/kubernetes/vendor/k8s.io/apimachinery/pkg/util/wait.WaitForWithContext({0x7fe0bc8, 0xc0000820c8}, 0xc003634ca8, 0x2fdb16a?) vendor/k8s.io/apimachinery/pkg/util/wait/wait.go:660 k8s.io/kubernetes/vendor/k8s.io/apimachinery/pkg/util/wait.poll({0x7fe0bc8, 0xc0000820c8}, 0xd0?, 0x2fd9d05?, 0x28?) vendor/k8s.io/apimachinery/pkg/util/wait/wait.go:596 k8s.io/kubernetes/vendor/k8s.io/apimachinery/pkg/util/wait.PollImmediateWithContext({0x7fe0bc8, 0xc0000820c8}, 0x0?, 0xc003c87c20?, 0x262a967?) vendor/k8s.io/apimachinery/pkg/util/wait/wait.go:528 k8s.io/kubernetes/vendor/k8s.io/apimachinery/pkg/util/wait.PollImmediate(0x754e980?, 0xc000dee5a0?, 0x766a5c9?) vendor/k8s.io/apimachinery/pkg/util/wait/wait.go:514 k8s.io/kubernetes/test/e2e/framework/service.TestReachableHTTPWithRetriableErrorCodes({0xc0022ac0f0, 0xe}, 0x7d7c, {0xae73300, 0x0, 0x0}, 0x1?) test/e2e/framework/service/util.go:46 k8s.io/kubernetes/test/e2e/framework/service.TestReachableHTTP(...) test/e2e/framework/service/util.go:29 > k8s.io/kubernetes/test/e2e/network.glob..func19.3() test/e2e/network/loadbalancer.go:261 k8s.io/kubernetes/vendor/github.com/onsi/ginkgo/v2/internal.extractBodyFunction.func3({0x2d591ce, 0xc001dea780}) vendor/github.com/onsi/ginkgo/v2/internal/node.go:449 k8s.io/kubernetes/vendor/github.com/onsi/ginkgo/v2/internal.(*Suite).runNode.func2() vendor/github.com/onsi/ginkgo/v2/internal/suite.go:750 k8s.io/kubernetes/vendor/github.com/onsi/ginkgo/v2/internal.(*Suite).runNode vendor/github.com/onsi/ginkgo/v2/internal/suite.go:738 ------------------------------ Nov 26 05:13:55.518: INFO: Poking "http://35.247.104.236:32124/echo?msg=hello" Nov 26 05:13:55.558: INFO: Poke("http://35.247.104.236:32124/echo?msg=hello"): Get "http://35.247.104.236:32124/echo?msg=hello": dial tcp 35.247.104.236:32124: connect: no route to host Nov 26 05:13:57.519: INFO: Poking "http://35.247.104.236:32124/echo?msg=hello" Nov 26 05:13:57.600: INFO: Poke("http://35.247.104.236:32124/echo?msg=hello"): success STEP: hitting the TCP service's LoadBalancer 11/26/22 05:13:57.6 Nov 26 05:13:57.601: INFO: Poking "http://35.185.230.152:81/echo?msg=hello" Nov 26 05:13:57.681: INFO: Poke("http://35.185.230.152:81/echo?msg=hello"): success STEP: changing TCP service back to type=ClusterIP 11/26/22 05:13:57.681 Nov 26 05:13:57.720: INFO: Unexpected error: <*errors.errorString | 0xc00133d8d0>: { s: "failed to get Service \"mutability-test\": Get \"https://35.227.179.144/api/v1/namespaces/loadbalancers-1625/services/mutability-test\": dial tcp 35.227.179.144:443: connect: connection refused", } Nov 26 05:13:57.720: FAIL: failed to get Service "mutability-test": Get "https://35.227.179.144/api/v1/namespaces/loadbalancers-1625/services/mutability-test": dial tcp 35.227.179.144:443: connect: connection refused Full Stack Trace k8s.io/kubernetes/test/e2e/network.glob..func19.3() test/e2e/network/loadbalancer.go:272 +0x17c5 [AfterEach] [sig-network] LoadBalancers test/e2e/framework/node/init/init.go:32 Nov 26 05:13:57.720: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready [AfterEach] [sig-network] LoadBalancers test/e2e/network/loadbalancer.go:71 Nov 26 05:13:57.760: INFO: Output of kubectl describe svc: Nov 26 05:13:57.760: INFO: Running '/workspace/github.com/containerd/containerd/kubernetes/platforms/linux/amd64/kubectl --server=https://35.227.179.144 --kubeconfig=/workspace/.kube/config --namespace=loadbalancers-1625 describe svc --namespace=loadbalancers-1625' Nov 26 05:13:57.880: INFO: rc: 1 Nov 26 05:13:57.881: INFO: [DeferCleanup (Each)] [sig-network] LoadBalancers test/e2e/framework/metrics/init/init.go:33 [DeferCleanup (Each)] [sig-network] LoadBalancers dump namespaces | framework.go:196 STEP: dump namespace information after failure 11/26/22 05:13:57.881 STEP: Collecting events from namespace "loadbalancers-1625". 11/26/22 05:13:57.881 Nov 26 05:13:57.920: INFO: Unexpected error: failed to list events in namespace "loadbalancers-1625": <*url.Error | 0xc0011f66f0>: { Op: "Get", URL: "https://35.227.179.144/api/v1/namespaces/loadbalancers-1625/events", Err: <*net.OpError | 0xc00297d770>{ Op: "dial", Net: "tcp", Source: nil, Addr: <*net.TCPAddr | 0xc002187470>{ IP: [0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 255, 255, 35, 227, 179, 144], Port: 443, Zone: "", }, Err: <*os.SyscallError | 0xc00446c8a0>{ Syscall: "connect", Err: <syscall.Errno>0x6f, }, }, } Nov 26 05:13:57.920: FAIL: failed to list events in namespace "loadbalancers-1625": Get "https://35.227.179.144/api/v1/namespaces/loadbalancers-1625/events": dial tcp 35.227.179.144:443: connect: connection refused Full Stack Trace k8s.io/kubernetes/test/e2e/framework/debug.dumpEventsInNamespace(0xc0022f85c0, {0xc003c03320, 0x12}) test/e2e/framework/debug/dump.go:44 +0x191 k8s.io/kubernetes/test/e2e/framework/debug.DumpAllNamespaceInfo({0x801de88, 0xc001f42820}, {0xc003c03320, 0x12}) test/e2e/framework/debug/dump.go:62 +0x8d k8s.io/kubernetes/test/e2e/framework/debug/init.init.0.func1.1(0xc0022f8650?, {0xc003c03320?, 0x7fa7740?}) test/e2e/framework/debug/init/init.go:34 +0x32 k8s.io/kubernetes/test/e2e/framework.(*Framework).dumpNamespaceInfo.func1() test/e2e/framework/framework.go:274 +0x6d k8s.io/kubernetes/test/e2e/framework.(*Framework).dumpNamespaceInfo(0xc000d0c4b0) test/e2e/framework/framework.go:271 +0x179 reflect.Value.call({0x6627cc0?, 0xc001698820?, 0xc00393cf50?}, {0x75b6e72, 0x4}, {0xae73300, 0x0, 0x0?}) /usr/local/go/src/reflect/value.go:584 +0x8c5 reflect.Value.Call({0x6627cc0?, 0xc001698820?, 0x7fadfa0?}, {0xae73300?, 0xc00393cf80?, 0x26225bd?}) /usr/local/go/src/reflect/value.go:368 +0xbc [DeferCleanup (Each)] [sig-network] LoadBalancers tear down framework | framework.go:193 STEP: Destroying namespace "loadbalancers-1625" for this suite. 11/26/22 05:13:57.921 Nov 26 05:13:57.960: FAIL: Couldn't delete ns: "loadbalancers-1625": Delete "https://35.227.179.144/api/v1/namespaces/loadbalancers-1625": dial tcp 35.227.179.144:443: connect: connection refused (&url.Error{Op:"Delete", URL:"https://35.227.179.144/api/v1/namespaces/loadbalancers-1625", Err:(*net.OpError)(0xc003924320)}) Full Stack Trace k8s.io/kubernetes/test/e2e/framework.(*Framework).AfterEach.func1() test/e2e/framework/framework.go:370 +0x4fe k8s.io/kubernetes/test/e2e/framework.(*Framework).AfterEach(0xc000d0c4b0) test/e2e/framework/framework.go:383 +0x1ca reflect.Value.call({0x6627cc0?, 0xc001698760?, 0xc0038a2fb0?}, {0x75b6e72, 0x4}, {0xae73300, 0x0, 0x0?}) /usr/local/go/src/reflect/value.go:584 +0x8c5 reflect.Value.Call({0x6627cc0?, 0xc001698760?, 0x0?}, {0xae73300?, 0x5?, 0xc003635098?}) /usr/local/go/src/reflect/value.go:368 +0xbc
Filter through log files | View test history on testgrid
go run hack/e2e.go -v --test --test_args='--ginkgo.focus=Kubernetes\se2e\ssuite\s\[It\]\s\[sig\-network\]\sLoadBalancers\sshould\sbe\sable\sto\schange\sthe\stype\sand\sports\sof\sa\sUDP\sservice\s\[Slow\]$'
test/e2e/network/service.go:604 k8s.io/kubernetes/test/e2e/network.testReachableUDP({0xc004c3de70, 0xe}, 0x7acb, 0x0?) test/e2e/network/service.go:604 +0x17b k8s.io/kubernetes/test/e2e/network.glob..func19.4() test/e2e/network/loadbalancer.go:390 +0xb25
[BeforeEach] [sig-network] LoadBalancers set up framework | framework.go:178 STEP: Creating a kubernetes client 11/26/22 05:23:40.069 Nov 26 05:23:40.069: INFO: >>> kubeConfig: /workspace/.kube/config STEP: Building a namespace api object, basename loadbalancers 11/26/22 05:23:40.071 STEP: Waiting for a default service account to be provisioned in namespace 11/26/22 05:23:40.366 STEP: Waiting for kube-root-ca.crt to be provisioned in namespace 11/26/22 05:23:40.464 [BeforeEach] [sig-network] LoadBalancers test/e2e/framework/metrics/init/init.go:31 [BeforeEach] [sig-network] LoadBalancers test/e2e/network/loadbalancer.go:65 [It] should be able to change the type and ports of a UDP service [Slow] test/e2e/network/loadbalancer.go:287 Nov 26 05:23:40.700: INFO: namespace for TCP test: loadbalancers-6791 STEP: creating a UDP service mutability-test with type=ClusterIP in namespace loadbalancers-6791 11/26/22 05:23:40.76 Nov 26 05:23:40.836: INFO: service port UDP: 80 STEP: creating a pod to be part of the UDP service mutability-test 11/26/22 05:23:40.836 Nov 26 05:23:40.903: INFO: Waiting up to 2m0s for 1 pods to be created Nov 26 05:23:40.978: INFO: Found all 1 pods Nov 26 05:23:40.978: INFO: Waiting up to 2m0s for 1 pods to be running and ready: [mutability-test-7ktz4] Nov 26 05:23:40.978: INFO: Waiting up to 2m0s for pod "mutability-test-7ktz4" in namespace "loadbalancers-6791" to be "running and ready" Nov 26 05:23:41.034: INFO: Pod "mutability-test-7ktz4": Phase="Pending", Reason="", readiness=false. Elapsed: 56.288605ms Nov 26 05:23:41.034: INFO: Error evaluating pod condition running and ready: want pod 'mutability-test-7ktz4' on 'bootstrap-e2e-minion-group-vhdj' to be 'Running' but was 'Pending' Nov 26 05:23:43.201: INFO: Pod "mutability-test-7ktz4": Phase="Pending", Reason="", readiness=false. Elapsed: 2.223360905s Nov 26 05:23:43.201: INFO: Error evaluating pod condition running and ready: want pod 'mutability-test-7ktz4' on 'bootstrap-e2e-minion-group-vhdj' to be 'Running' but was 'Pending' Nov 26 05:23:45.100: INFO: Pod "mutability-test-7ktz4": Phase="Running", Reason="", readiness=true. Elapsed: 4.122679584s Nov 26 05:23:45.100: INFO: Pod "mutability-test-7ktz4" satisfied condition "running and ready" Nov 26 05:23:45.101: INFO: Wanted all 1 pods to be running and ready. Result: true. Pods: [mutability-test-7ktz4] STEP: changing the UDP service to type=NodePort 11/26/22 05:23:45.101 Nov 26 05:23:45.249: INFO: UDP node port: 31435 STEP: hitting the UDP service's NodePort 11/26/22 05:23:45.249 Nov 26 05:23:45.249: INFO: Poking udp://35.247.104.236:31435 Nov 26 05:23:45.290: INFO: Poke("udp://35.247.104.236:31435"): read udp 10.60.29.184:42588->35.247.104.236:31435: read: connection refused Nov 26 05:23:47.290: INFO: Poking udp://35.247.104.236:31435 Nov 26 05:23:47.329: INFO: Poke("udp://35.247.104.236:31435"): read udp 10.60.29.184:54376->35.247.104.236:31435: read: connection refused Nov 26 05:23:49.290: INFO: Poking udp://35.247.104.236:31435 Nov 26 05:23:49.329: INFO: Poke("udp://35.247.104.236:31435"): read udp 10.60.29.184:40521->35.247.104.236:31435: read: connection refused Nov 26 05:23:51.290: INFO: Poking udp://35.247.104.236:31435 Nov 26 05:23:51.329: INFO: Poke("udp://35.247.104.236:31435"): read udp 10.60.29.184:59605->35.247.104.236:31435: read: connection refused Nov 26 05:23:53.290: INFO: Poking udp://35.247.104.236:31435 Nov 26 05:23:53.329: INFO: Poke("udp://35.247.104.236:31435"): read udp 10.60.29.184:59601->35.247.104.236:31435: read: connection refused Nov 26 05:23:55.290: INFO: Poking udp://35.247.104.236:31435 Nov 26 05:23:55.331: INFO: Poke("udp://35.247.104.236:31435"): success STEP: creating a static load balancer IP 11/26/22 05:23:55.331 Nov 26 05:23:57.366: INFO: Allocated static load balancer IP: 34.168.22.95 STEP: changing the UDP service to type=LoadBalancer 11/26/22 05:23:57.366 STEP: demoting the static IP to ephemeral 11/26/22 05:23:57.451 STEP: waiting for the UDP service to have a load balancer 11/26/22 05:23:58.983 Nov 26 05:23:58.983: INFO: Waiting up to 15m0s for service "mutability-test" to have a LoadBalancer Nov 26 05:26:43.065: INFO: Retrying .... error trying to get Service mutability-test: Get "https://35.227.179.144/api/v1/namespaces/loadbalancers-6791/services/mutability-test": stream error: stream ID 557; INTERNAL_ERROR; received from peer ------------------------------ Progress Report for Ginkgo Process #6 Automatically polling progress: [sig-network] LoadBalancers should be able to change the type and ports of a UDP service [Slow] (Spec Runtime: 5m0.563s) test/e2e/network/loadbalancer.go:287 In [It] (Node Runtime: 5m0s) test/e2e/network/loadbalancer.go:287 At [By Step] waiting for the UDP service to have a load balancer (Step Runtime: 4m41.65s) test/e2e/network/loadbalancer.go:379 Spec Goroutine goroutine 1768 [select] k8s.io/kubernetes/vendor/k8s.io/apimachinery/pkg/util/wait.WaitForWithContext({0x7fe0bc8, 0xc0000820c8}, 0xc000e32cc0, 0x2fdb16a?) vendor/k8s.io/apimachinery/pkg/util/wait/wait.go:660 k8s.io/kubernetes/vendor/k8s.io/apimachinery/pkg/util/wait.poll({0x7fe0bc8, 0xc0000820c8}, 0x68?, 0x2fd9d05?, 0x20?) vendor/k8s.io/apimachinery/pkg/util/wait/wait.go:596 k8s.io/kubernetes/vendor/k8s.io/apimachinery/pkg/util/wait.PollImmediateWithContext({0x7fe0bc8, 0xc0000820c8}, 0xc0031e9b60?, 0xc003475bb8?, 0x262a967?) vendor/k8s.io/apimachinery/pkg/util/wait/wait.go:528 k8s.io/kubernetes/vendor/k8s.io/apimachinery/pkg/util/wait.PollImmediate(0xc000dd2130?, 0x7fa7740?, 0xc000202b80?) vendor/k8s.io/apimachinery/pkg/util/wait/wait.go:514 k8s.io/kubernetes/test/e2e/framework/service.(*TestJig).waitForCondition(0xc003b567d0, 0x4?, {0x7600fe2, 0x14}, 0x7895b68) test/e2e/framework/service/jig.go:631 k8s.io/kubernetes/test/e2e/framework/service.(*TestJig).WaitForLoadBalancer(0xc003b567d0, 0x33?) test/e2e/framework/service/jig.go:582 > k8s.io/kubernetes/test/e2e/network.glob..func19.4() test/e2e/network/loadbalancer.go:381 k8s.io/kubernetes/vendor/github.com/onsi/ginkgo/v2/internal.extractBodyFunction.func3({0x2d591ce, 0xc0056e0000}) vendor/github.com/onsi/ginkgo/v2/internal/node.go:449 k8s.io/kubernetes/vendor/github.com/onsi/ginkgo/v2/internal.(*Suite).runNode.func2() vendor/github.com/onsi/ginkgo/v2/internal/suite.go:750 k8s.io/kubernetes/vendor/github.com/onsi/ginkgo/v2/internal.(*Suite).runNode vendor/github.com/onsi/ginkgo/v2/internal/suite.go:738 ------------------------------ ------------------------------ Progress Report for Ginkgo Process #6 Automatically polling progress: [sig-network] LoadBalancers should be able to change the type and ports of a UDP service [Slow] (Spec Runtime: 5m20.565s) test/e2e/network/loadbalancer.go:287 In [It] (Node Runtime: 5m20.002s) test/e2e/network/loadbalancer.go:287 At [By Step] waiting for the UDP service to have a load balancer (Step Runtime: 5m1.651s) test/e2e/network/loadbalancer.go:379 Spec Goroutine goroutine 1768 [select] k8s.io/kubernetes/vendor/k8s.io/apimachinery/pkg/util/wait.WaitForWithContext({0x7fe0bc8, 0xc0000820c8}, 0xc000e32cc0, 0x2fdb16a?) vendor/k8s.io/apimachinery/pkg/util/wait/wait.go:660 k8s.io/kubernetes/vendor/k8s.io/apimachinery/pkg/util/wait.poll({0x7fe0bc8, 0xc0000820c8}, 0x68?, 0x2fd9d05?, 0x20?) vendor/k8s.io/apimachinery/pkg/util/wait/wait.go:596 k8s.io/kubernetes/vendor/k8s.io/apimachinery/pkg/util/wait.PollImmediateWithContext({0x7fe0bc8, 0xc0000820c8}, 0xc0031e9b60?, 0xc003475bb8?, 0x262a967?) vendor/k8s.io/apimachinery/pkg/util/wait/wait.go:528 k8s.io/kubernetes/vendor/k8s.io/apimachinery/pkg/util/wait.PollImmediate(0xc000dd2130?, 0x7fa7740?, 0xc000202b80?) vendor/k8s.io/apimachinery/pkg/util/wait/wait.go:514 k8s.io/kubernetes/test/e2e/framework/service.(*TestJig).waitForCondition(0xc003b567d0, 0x4?, {0x7600fe2, 0x14}, 0x7895b68) test/e2e/framework/service/jig.go:631 k8s.io/kubernetes/test/e2e/framework/service.(*TestJig).WaitForLoadBalancer(0xc003b567d0, 0x33?) test/e2e/framework/service/jig.go:582 > k8s.io/kubernetes/test/e2e/network.glob..func19.4() test/e2e/network/loadbalancer.go:381 k8s.io/kubernetes/vendor/github.com/onsi/ginkgo/v2/internal.extractBodyFunction.func3({0x2d591ce, 0xc0056e0000}) vendor/github.com/onsi/ginkgo/v2/internal/node.go:449 k8s.io/kubernetes/vendor/github.com/onsi/ginkgo/v2/internal.(*Suite).runNode.func2() vendor/github.com/onsi/ginkgo/v2/internal/suite.go:750 k8s.io/kubernetes/vendor/github.com/onsi/ginkgo/v2/internal.(*Suite).runNode vendor/github.com/onsi/ginkgo/v2/internal/suite.go:738 ------------------------------ ------------------------------ Progress Report for Ginkgo Process #6 Automatically polling progress: [sig-network] LoadBalancers should be able to change the type and ports of a UDP service [Slow] (Spec Runtime: 5m40.567s) test/e2e/network/loadbalancer.go:287 In [It] (Node Runtime: 5m40.004s) test/e2e/network/loadbalancer.go:287 At [By Step] waiting for the UDP service to have a load balancer (Step Runtime: 5m21.653s) test/e2e/network/loadbalancer.go:379 Spec Goroutine goroutine 1768 [select] k8s.io/kubernetes/vendor/k8s.io/apimachinery/pkg/util/wait.WaitForWithContext({0x7fe0bc8, 0xc0000820c8}, 0xc000e32cc0, 0x2fdb16a?) vendor/k8s.io/apimachinery/pkg/util/wait/wait.go:660 k8s.io/kubernetes/vendor/k8s.io/apimachinery/pkg/util/wait.poll({0x7fe0bc8, 0xc0000820c8}, 0x68?, 0x2fd9d05?, 0x20?) vendor/k8s.io/apimachinery/pkg/util/wait/wait.go:596 k8s.io/kubernetes/vendor/k8s.io/apimachinery/pkg/util/wait.PollImmediateWithContext({0x7fe0bc8, 0xc0000820c8}, 0xc0031e9b60?, 0xc003475bb8?, 0x262a967?) vendor/k8s.io/apimachinery/pkg/util/wait/wait.go:528 k8s.io/kubernetes/vendor/k8s.io/apimachinery/pkg/util/wait.PollImmediate(0xc000dd2130?, 0x7fa7740?, 0xc000202b80?) vendor/k8s.io/apimachinery/pkg/util/wait/wait.go:514 k8s.io/kubernetes/test/e2e/framework/service.(*TestJig).waitForCondition(0xc003b567d0, 0x4?, {0x7600fe2, 0x14}, 0x7895b68) test/e2e/framework/service/jig.go:631 k8s.io/kubernetes/test/e2e/framework/service.(*TestJig).WaitForLoadBalancer(0xc003b567d0, 0x33?) test/e2e/framework/service/jig.go:582 > k8s.io/kubernetes/test/e2e/network.glob..func19.4() test/e2e/network/loadbalancer.go:381 k8s.io/kubernetes/vendor/github.com/onsi/ginkgo/v2/internal.extractBodyFunction.func3({0x2d591ce, 0xc0056e0000}) vendor/github.com/onsi/ginkgo/v2/internal/node.go:449 k8s.io/kubernetes/vendor/github.com/onsi/ginkgo/v2/internal.(*Suite).runNode.func2() vendor/github.com/onsi/ginkgo/v2/internal/suite.go:750 k8s.io/kubernetes/vendor/github.com/onsi/ginkgo/v2/internal.(*Suite).runNode vendor/github.com/onsi/ginkgo/v2/internal/suite.go:738 ------------------------------ ------------------------------ Progress Report for Ginkgo Process #6 Automatically polling progress: [sig-network] LoadBalancers should be able to change the type and ports of a UDP service [Slow] (Spec Runtime: 6m0.569s) test/e2e/network/loadbalancer.go:287 In [It] (Node Runtime: 6m0.006s) test/e2e/network/loadbalancer.go:287 At [By Step] waiting for the UDP service to have a load balancer (Step Runtime: 5m41.656s) test/e2e/network/loadbalancer.go:379 Spec Goroutine goroutine 1768 [select] k8s.io/kubernetes/vendor/k8s.io/apimachinery/pkg/util/wait.WaitForWithContext({0x7fe0bc8, 0xc0000820c8}, 0xc000e32cc0, 0x2fdb16a?) vendor/k8s.io/apimachinery/pkg/util/wait/wait.go:660 k8s.io/kubernetes/vendor/k8s.io/apimachinery/pkg/util/wait.poll({0x7fe0bc8, 0xc0000820c8}, 0x68?, 0x2fd9d05?, 0x20?) vendor/k8s.io/apimachinery/pkg/util/wait/wait.go:596 k8s.io/kubernetes/vendor/k8s.io/apimachinery/pkg/util/wait.PollImmediateWithContext({0x7fe0bc8, 0xc0000820c8}, 0xc0031e9b60?, 0xc003475bb8?, 0x262a967?) vendor/k8s.io/apimachinery/pkg/util/wait/wait.go:528 k8s.io/kubernetes/vendor/k8s.io/apimachinery/pkg/util/wait.PollImmediate(0xc000dd2130?, 0x7fa7740?, 0xc000202b80?) vendor/k8s.io/apimachinery/pkg/util/wait/wait.go:514 k8s.io/kubernetes/test/e2e/framework/service.(*TestJig).waitForCondition(0xc003b567d0, 0x4?, {0x7600fe2, 0x14}, 0x7895b68) test/e2e/framework/service/jig.go:631 k8s.io/kubernetes/test/e2e/framework/service.(*TestJig).WaitForLoadBalancer(0xc003b567d0, 0x33?) test/e2e/framework/service/jig.go:582 > k8s.io/kubernetes/test/e2e/network.glob..func19.4() test/e2e/network/loadbalancer.go:381 k8s.io/kubernetes/vendor/github.com/onsi/ginkgo/v2/internal.extractBodyFunction.func3({0x2d591ce, 0xc0056e0000}) vendor/github.com/onsi/ginkgo/v2/internal/node.go:449 k8s.io/kubernetes/vendor/github.com/onsi/ginkgo/v2/internal.(*Suite).runNode.func2() vendor/github.com/onsi/ginkgo/v2/internal/suite.go:750 k8s.io/kubernetes/vendor/github.com/onsi/ginkgo/v2/internal.(*Suite).runNode vendor/github.com/onsi/ginkgo/v2/internal/suite.go:738 ------------------------------ ------------------------------ Progress Report for Ginkgo Process #6 Automatically polling progress: [sig-network] LoadBalancers should be able to change the type and ports of a UDP service [Slow] (Spec Runtime: 6m20.572s) test/e2e/network/loadbalancer.go:287 In [It] (Node Runtime: 6m20.01s) test/e2e/network/loadbalancer.go:287 At [By Step] waiting for the UDP service to have a load balancer (Step Runtime: 6m1.659s) test/e2e/network/loadbalancer.go:379 Spec Goroutine goroutine 1768 [select] k8s.io/kubernetes/vendor/k8s.io/apimachinery/pkg/util/wait.WaitForWithContext({0x7fe0bc8, 0xc0000820c8}, 0xc000e32cc0, 0x2fdb16a?) vendor/k8s.io/apimachinery/pkg/util/wait/wait.go:660 k8s.io/kubernetes/vendor/k8s.io/apimachinery/pkg/util/wait.poll({0x7fe0bc8, 0xc0000820c8}, 0x68?, 0x2fd9d05?, 0x20?) vendor/k8s.io/apimachinery/pkg/util/wait/wait.go:596 k8s.io/kubernetes/vendor/k8s.io/apimachinery/pkg/util/wait.PollImmediateWithContext({0x7fe0bc8, 0xc0000820c8}, 0xc0031e9b60?, 0xc003475bb8?, 0x262a967?) vendor/k8s.io/apimachinery/pkg/util/wait/wait.go:528 k8s.io/kubernetes/vendor/k8s.io/apimachinery/pkg/util/wait.PollImmediate(0xc000dd2130?, 0x7fa7740?, 0xc000202b80?) vendor/k8s.io/apimachinery/pkg/util/wait/wait.go:514 k8s.io/kubernetes/test/e2e/framework/service.(*TestJig).waitForCondition(0xc003b567d0, 0x4?, {0x7600fe2, 0x14}, 0x7895b68) test/e2e/framework/service/jig.go:631 k8s.io/kubernetes/test/e2e/framework/service.(*TestJig).WaitForLoadBalancer(0xc003b567d0, 0x33?) test/e2e/framework/service/jig.go:582 > k8s.io/kubernetes/test/e2e/network.glob..func19.4() test/e2e/network/loadbalancer.go:381 k8s.io/kubernetes/vendor/github.com/onsi/ginkgo/v2/internal.extractBodyFunction.func3({0x2d591ce, 0xc0056e0000}) vendor/github.com/onsi/ginkgo/v2/internal/node.go:449 k8s.io/kubernetes/vendor/github.com/onsi/ginkgo/v2/internal.(*Suite).runNode.func2() vendor/github.com/onsi/ginkgo/v2/internal/suite.go:750 k8s.io/kubernetes/vendor/github.com/onsi/ginkgo/v2/internal.(*Suite).runNode vendor/github.com/onsi/ginkgo/v2/internal/suite.go:738 ------------------------------ ------------------------------ Progress Report for Ginkgo Process #6 Automatically polling progress: [sig-network] LoadBalancers should be able to change the type and ports of a UDP service [Slow] (Spec Runtime: 6m40.574s) test/e2e/network/loadbalancer.go:287 In [It] (Node Runtime: 6m40.011s) test/e2e/network/loadbalancer.go:287 At [By Step] waiting for the UDP service to have a load balancer (Step Runtime: 6m21.66s) test/e2e/network/loadbalancer.go:379 Spec Goroutine goroutine 1768 [select] k8s.io/kubernetes/vendor/k8s.io/apimachinery/pkg/util/wait.WaitForWithContext({0x7fe0bc8, 0xc0000820c8}, 0xc000e32cc0, 0x2fdb16a?) vendor/k8s.io/apimachinery/pkg/util/wait/wait.go:660 k8s.io/kubernetes/vendor/k8s.io/apimachinery/pkg/util/wait.poll({0x7fe0bc8, 0xc0000820c8}, 0x68?, 0x2fd9d05?, 0x20?) vendor/k8s.io/apimachinery/pkg/util/wait/wait.go:596 k8s.io/kubernetes/vendor/k8s.io/apimachinery/pkg/util/wait.PollImmediateWithContext({0x7fe0bc8, 0xc0000820c8}, 0xc0031e9b60?, 0xc003475bb8?, 0x262a967?) vendor/k8s.io/apimachinery/pkg/util/wait/wait.go:528 k8s.io/kubernetes/vendor/k8s.io/apimachinery/pkg/util/wait.PollImmediate(0xc000dd2130?, 0x7fa7740?, 0xc000202b80?) vendor/k8s.io/apimachinery/pkg/util/wait/wait.go:514 k8s.io/kubernetes/test/e2e/framework/service.(*TestJig).waitForCondition(0xc003b567d0, 0x4?, {0x7600fe2, 0x14}, 0x7895b68) test/e2e/framework/service/jig.go:631 k8s.io/kubernetes/test/e2e/framework/service.(*TestJig).WaitForLoadBalancer(0xc003b567d0, 0x33?) test/e2e/framework/service/jig.go:582 > k8s.io/kubernetes/test/e2e/network.glob..func19.4() test/e2e/network/loadbalancer.go:381 k8s.io/kubernetes/vendor/github.com/onsi/ginkgo/v2/internal.extractBodyFunction.func3({0x2d591ce, 0xc0056e0000}) vendor/github.com/onsi/ginkgo/v2/internal/node.go:449 k8s.io/kubernetes/vendor/github.com/onsi/ginkgo/v2/internal.(*Suite).runNode.func2() vendor/github.com/onsi/ginkgo/v2/internal/suite.go:750 k8s.io/kubernetes/vendor/github.com/onsi/ginkgo/v2/internal.(*Suite).runNode vendor/github.com/onsi/ginkgo/v2/internal/suite.go:738 ------------------------------ ------------------------------ Progress Report for Ginkgo Process #6 Automatically polling progress: [sig-network] LoadBalancers should be able to change the type and ports of a UDP service [Slow] (Spec Runtime: 7m0.583s) test/e2e/network/loadbalancer.go:287 In [It] (Node Runtime: 7m0.02s) test/e2e/network/loadbalancer.go:287 At [By Step] waiting for the UDP service to have a load balancer (Step Runtime: 6m41.67s) test/e2e/network/loadbalancer.go:379 Spec Goroutine goroutine 1768 [select] k8s.io/kubernetes/vendor/k8s.io/apimachinery/pkg/util/wait.WaitForWithContext({0x7fe0bc8, 0xc0000820c8}, 0xc000e32cc0, 0x2fdb16a?) vendor/k8s.io/apimachinery/pkg/util/wait/wait.go:660 k8s.io/kubernetes/vendor/k8s.io/apimachinery/pkg/util/wait.poll({0x7fe0bc8, 0xc0000820c8}, 0x68?, 0x2fd9d05?, 0x20?) vendor/k8s.io/apimachinery/pkg/util/wait/wait.go:596 k8s.io/kubernetes/vendor/k8s.io/apimachinery/pkg/util/wait.PollImmediateWithContext({0x7fe0bc8, 0xc0000820c8}, 0xc0031e9b60?, 0xc003475bb8?, 0x262a967?) vendor/k8s.io/apimachinery/pkg/util/wait/wait.go:528 k8s.io/kubernetes/vendor/k8s.io/apimachinery/pkg/util/wait.PollImmediate(0xc000dd2130?, 0x7fa7740?, 0xc000202b80?) vendor/k8s.io/apimachinery/pkg/util/wait/wait.go:514 k8s.io/kubernetes/test/e2e/framework/service.(*TestJig).waitForCondition(0xc003b567d0, 0x4?, {0x7600fe2, 0x14}, 0x7895b68) test/e2e/framework/service/jig.go:631 k8s.io/kubernetes/test/e2e/framework/service.(*TestJig).WaitForLoadBalancer(0xc003b567d0, 0x33?) test/e2e/framework/service/jig.go:582 > k8s.io/kubernetes/test/e2e/network.glob..func19.4() test/e2e/network/loadbalancer.go:381 k8s.io/kubernetes/vendor/github.com/onsi/ginkgo/v2/internal.extractBodyFunction.func3({0x2d591ce, 0xc0056e0000}) vendor/github.com/onsi/ginkgo/v2/internal/node.go:449 k8s.io/kubernetes/vendor/github.com/onsi/ginkgo/v2/internal.(*Suite).runNode.func2() vendor/github.com/onsi/ginkgo/v2/internal/suite.go:750 k8s.io/kubernetes/vendor/github.com/onsi/ginkgo/v2/internal.(*Suite).runNode vendor/github.com/onsi/ginkgo/v2/internal/suite.go:738 ------------------------------ ------------------------------ Progress Report for Ginkgo Process #6 Automatically polling progress: [sig-network] LoadBalancers should be able to change the type and ports of a UDP service [Slow] (Spec Runtime: 7m20.59s) test/e2e/network/loadbalancer.go:287 In [It] (Node Runtime: 7m20.027s) test/e2e/network/loadbalancer.go:287 At [By Step] waiting for the UDP service to have a load balancer (Step Runtime: 7m1.676s) test/e2e/network/loadbalancer.go:379 Spec Goroutine goroutine 1768 [select] k8s.io/kubernetes/vendor/k8s.io/apimachinery/pkg/util/wait.WaitForWithContext({0x7fe0bc8, 0xc0000820c8}, 0xc000e32cc0, 0x2fdb16a?) vendor/k8s.io/apimachinery/pkg/util/wait/wait.go:660 k8s.io/kubernetes/vendor/k8s.io/apimachinery/pkg/util/wait.poll({0x7fe0bc8, 0xc0000820c8}, 0x68?, 0x2fd9d05?, 0x20?) vendor/k8s.io/apimachinery/pkg/util/wait/wait.go:596 k8s.io/kubernetes/vendor/k8s.io/apimachinery/pkg/util/wait.PollImmediateWithContext({0x7fe0bc8, 0xc0000820c8}, 0xc0031e9b60?, 0xc003475bb8?, 0x262a967?) vendor/k8s.io/apimachinery/pkg/util/wait/wait.go:528 k8s.io/kubernetes/vendor/k8s.io/apimachinery/pkg/util/wait.PollImmediate(0xc000dd2130?, 0x7fa7740?, 0xc000202b80?) vendor/k8s.io/apimachinery/pkg/util/wait/wait.go:514 k8s.io/kubernetes/test/e2e/framework/service.(*TestJig).waitForCondition(0xc003b567d0, 0x4?, {0x7600fe2, 0x14}, 0x7895b68) test/e2e/framework/service/jig.go:631 k8s.io/kubernetes/test/e2e/framework/service.(*TestJig).WaitForLoadBalancer(0xc003b567d0, 0x33?) test/e2e/framework/service/jig.go:582 > k8s.io/kubernetes/test/e2e/network.glob..func19.4() test/e2e/network/loadbalancer.go:381 k8s.io/kubernetes/vendor/github.com/onsi/ginkgo/v2/internal.extractBodyFunction.func3({0x2d591ce, 0xc0056e0000}) vendor/github.com/onsi/ginkgo/v2/internal/node.go:449 k8s.io/kubernetes/vendor/github.com/onsi/ginkgo/v2/internal.(*Suite).runNode.func2() vendor/github.com/onsi/ginkgo/v2/internal/suite.go:750 k8s.io/kubernetes/vendor/github.com/onsi/ginkgo/v2/internal.(*Suite).runNode vendor/github.com/onsi/ginkgo/v2/internal/suite.go:738 ------------------------------ ------------------------------ Progress Report for Ginkgo Process #6 Automatically polling progress: [sig-network] LoadBalancers should be able to change the type and ports of a UDP service [Slow] (Spec Runtime: 7m40.593s) test/e2e/network/loadbalancer.go:287 In [It] (Node Runtime: 7m40.03s) test/e2e/network/loadbalancer.go:287 At [By Step] waiting for the UDP service to have a load balancer (Step Runtime: 7m21.679s) test/e2e/network/loadbalancer.go:379 Spec Goroutine goroutine 1768 [select] k8s.io/kubernetes/vendor/k8s.io/apimachinery/pkg/util/wait.WaitForWithContext({0x7fe0bc8, 0xc0000820c8}, 0xc000e32cc0, 0x2fdb16a?) vendor/k8s.io/apimachinery/pkg/util/wait/wait.go:660 k8s.io/kubernetes/vendor/k8s.io/apimachinery/pkg/util/wait.poll({0x7fe0bc8, 0xc0000820c8}, 0x68?, 0x2fd9d05?, 0x20?) vendor/k8s.io/apimachinery/pkg/util/wait/wait.go:596 k8s.io/kubernetes/vendor/k8s.io/apimachinery/pkg/util/wait.PollImmediateWithContext({0x7fe0bc8, 0xc0000820c8}, 0xc0031e9b60?, 0xc003475bb8?, 0x262a967?) vendor/k8s.io/apimachinery/pkg/util/wait/wait.go:528 k8s.io/kubernetes/vendor/k8s.io/apimachinery/pkg/util/wait.PollImmediate(0xc000dd2130?, 0x7fa7740?, 0xc000202b80?) vendor/k8s.io/apimachinery/pkg/util/wait/wait.go:514 k8s.io/kubernetes/test/e2e/framework/service.(*TestJig).waitForCondition(0xc003b567d0, 0x4?, {0x7600fe2, 0x14}, 0x7895b68) test/e2e/framework/service/jig.go:631 k8s.io/kubernetes/test/e2e/framework/service.(*TestJig).WaitForLoadBalancer(0xc003b567d0, 0x33?) test/e2e/framework/service/jig.go:582 > k8s.io/kubernetes/test/e2e/network.glob..func19.4() test/e2e/network/loadbalancer.go:381 k8s.io/kubernetes/vendor/github.com/onsi/ginkgo/v2/internal.extractBodyFunction.func3({0x2d591ce, 0xc0056e0000}) vendor/github.com/onsi/ginkgo/v2/internal/node.go:449 k8s.io/kubernetes/vendor/github.com/onsi/ginkgo/v2/internal.(*Suite).runNode.func2() vendor/github.com/onsi/ginkgo/v2/internal/suite.go:750 k8s.io/kubernetes/vendor/github.com/onsi/ginkgo/v2/internal.(*Suite).runNode vendor/github.com/onsi/ginkgo/v2/internal/suite.go:738 ------------------------------ ------------------------------ Progress Report for Ginkgo Process #6 Automatically polling progress: [sig-network] LoadBalancers should be able to change the type and ports of a UDP service [Slow] (Spec Runtime: 8m0.597s) test/e2e/network/loadbalancer.go:287 In [It] (Node Runtime: 8m0.034s) test/e2e/network/loadbalancer.go:287 At [By Step] waiting for the UDP service to have a load balancer (Step Runtime: 7m41.684s) test/e2e/network/loadbalancer.go:379 Spec Goroutine goroutine 1768 [select] k8s.io/kubernetes/vendor/k8s.io/apimachinery/pkg/util/wait.WaitForWithContext({0x7fe0bc8, 0xc0000820c8}, 0xc000e32cc0, 0x2fdb16a?) vendor/k8s.io/apimachinery/pkg/util/wait/wait.go:660 k8s.io/kubernetes/vendor/k8s.io/apimachinery/pkg/util/wait.poll({0x7fe0bc8, 0xc0000820c8}, 0x68?, 0x2fd9d05?, 0x20?) vendor/k8s.io/apimachinery/pkg/util/wait/wait.go:596 k8s.io/kubernetes/vendor/k8s.io/apimachinery/pkg/util/wait.PollImmediateWithContext({0x7fe0bc8, 0xc0000820c8}, 0xc0031e9b60?, 0xc003475bb8?, 0x262a967?) vendor/k8s.io/apimachinery/pkg/util/wait/wait.go:528 k8s.io/kubernetes/vendor/k8s.io/apimachinery/pkg/util/wait.PollImmediate(0xc000dd2130?, 0x7fa7740?, 0xc000202b80?) vendor/k8s.io/apimachinery/pkg/util/wait/wait.go:514 k8s.io/kubernetes/test/e2e/framework/service.(*TestJig).waitForCondition(0xc003b567d0, 0x4?, {0x7600fe2, 0x14}, 0x7895b68) test/e2e/framework/service/jig.go:631 k8s.io/kubernetes/test/e2e/framework/service.(*TestJig).WaitForLoadBalancer(0xc003b567d0, 0x33?) test/e2e/framework/service/jig.go:582 > k8s.io/kubernetes/test/e2e/network.glob..func19.4() test/e2e/network/loadbalancer.go:381 k8s.io/kubernetes/vendor/github.com/onsi/ginkgo/v2/internal.extractBodyFunction.func3({0x2d591ce, 0xc0056e0000}) vendor/github.com/onsi/ginkgo/v2/internal/node.go:449 k8s.io/kubernetes/vendor/github.com/onsi/ginkgo/v2/internal.(*Suite).runNode.func2() vendor/github.com/onsi/ginkgo/v2/internal/suite.go:750 k8s.io/kubernetes/vendor/github.com/onsi/ginkgo/v2/internal.(*Suite).runNode vendor/github.com/onsi/ginkgo/v2/internal/suite.go:738 ------------------------------ Nov 26 05:31:45.063: INFO: Retrying .... error trying to get Service mutability-test: Get "https://35.227.179.144/api/v1/namespaces/loadbalancers-6791/services/mutability-test": dial tcp 35.227.179.144:443: connect: connection refused Nov 26 05:31:47.064: INFO: Retrying .... error trying to get Service mutability-test: Get "https://35.227.179.144/api/v1/namespaces/loadbalancers-6791/services/mutability-test": dial tcp 35.227.179.144:443: connect: connection refused Nov 26 05:31:49.063: INFO: Retrying .... error trying to get Service mutability-test: Get "https://35.227.179.144/api/v1/namespaces/loadbalancers-6791/services/mutability-test": dial tcp 35.227.179.144:443: connect: connection refused Nov 26 05:31:51.063: INFO: Retrying .... error trying to get Service mutability-test: Get "https://35.227.179.144/api/v1/namespaces/loadbalancers-6791/services/mutability-test": dial tcp 35.227.179.144:443: connect: connection refused Nov 26 05:31:53.063: INFO: Retrying .... error trying to get Service mutability-test: Get "https://35.227.179.144/api/v1/namespaces/loadbalancers-6791/services/mutability-test": dial tcp 35.227.179.144:443: connect: connection refused Nov 26 05:31:55.063: INFO: Retrying .... error trying to get Service mutability-test: Get "https://35.227.179.144/api/v1/namespaces/loadbalancers-6791/services/mutability-test": dial tcp 35.227.179.144:443: connect: connection refused Nov 26 05:31:57.063: INFO: Retrying .... error trying to get Service mutability-test: Get "https://35.227.179.144/api/v1/namespaces/loadbalancers-6791/services/mutability-test": dial tcp 35.227.179.144:443: connect: connection refused Nov 26 05:31:59.063: INFO: Retrying .... error trying to get Service mutability-test: Get "https://35.227.179.144/api/v1/namespaces/loadbalancers-6791/services/mutability-test": dial tcp 35.227.179.144:443: connect: connection refused ------------------------------ Progress Report for Ginkgo Process #6 Automatically polling progress: [sig-network] LoadBalancers should be able to change the type and ports of a UDP service [Slow] (Spec Runtime: 8m20.604s) test/e2e/network/loadbalancer.go:287 In [It] (Node Runtime: 8m20.041s) test/e2e/network/loadbalancer.go:287 At [By Step] waiting for the UDP service to have a load balancer (Step Runtime: 8m1.69s) test/e2e/network/loadbalancer.go:379 Spec Goroutine goroutine 1768 [select] k8s.io/kubernetes/vendor/k8s.io/apimachinery/pkg/util/wait.WaitForWithContext({0x7fe0bc8, 0xc0000820c8}, 0xc000e32cc0, 0x2fdb16a?) vendor/k8s.io/apimachinery/pkg/util/wait/wait.go:660 k8s.io/kubernetes/vendor/k8s.io/apimachinery/pkg/util/wait.poll({0x7fe0bc8, 0xc0000820c8}, 0x68?, 0x2fd9d05?, 0x20?) vendor/k8s.io/apimachinery/pkg/util/wait/wait.go:596 k8s.io/kubernetes/vendor/k8s.io/apimachinery/pkg/util/wait.PollImmediateWithContext({0x7fe0bc8, 0xc0000820c8}, 0xc0031e9b60?, 0xc003475bb8?, 0x262a967?) vendor/k8s.io/apimachinery/pkg/util/wait/wait.go:528 k8s.io/kubernetes/vendor/k8s.io/apimachinery/pkg/util/wait.PollImmediate(0xc000dd2130?, 0x7fa7740?, 0xc000202b80?) vendor/k8s.io/apimachinery/pkg/util/wait/wait.go:514 k8s.io/kubernetes/test/e2e/framework/service.(*TestJig).waitForCondition(0xc003b567d0, 0x4?, {0x7600fe2, 0x14}, 0x7895b68) test/e2e/framework/service/jig.go:631 k8s.io/kubernetes/test/e2e/framework/service.(*TestJig).WaitForLoadBalancer(0xc003b567d0, 0x33?) test/e2e/framework/service/jig.go:582 > k8s.io/kubernetes/test/e2e/network.glob..func19.4() test/e2e/network/loadbalancer.go:381 k8s.io/kubernetes/vendor/github.com/onsi/ginkgo/v2/internal.extractBodyFunction.func3({0x2d591ce, 0xc0056e0000}) vendor/github.com/onsi/ginkgo/v2/internal/node.go:449 k8s.io/kubernetes/vendor/github.com/onsi/ginkgo/v2/internal.(*Suite).runNode.func2() vendor/github.com/onsi/ginkgo/v2/internal/suite.go:750 k8s.io/kubernetes/vendor/github.com/onsi/ginkgo/v2/internal.(*Suite).runNode vendor/github.com/onsi/ginkgo/v2/internal/suite.go:738 ------------------------------ Nov 26 05:32:01.063: INFO: Retrying .... error trying to get Service mutability-test: Get "https://35.227.179.144/api/v1/namespaces/loadbalancers-6791/services/mutability-test": dial tcp 35.227.179.144:443: connect: connection refused Nov 26 05:32:03.063: INFO: Retrying .... error trying to get Service mutability-test: Get "https://35.227.179.144/api/v1/namespaces/loadbalancers-6791/services/mutability-test": dial tcp 35.227.179.144:443: connect: connection refused Nov 26 05:32:05.064: INFO: Retrying .... error trying to get Service mutability-test: Get "https://35.227.179.144/api/v1/namespaces/loadbalancers-6791/services/mutability-test": dial tcp 35.227.179.144:443: connect: connection refused Nov 26 05:32:07.063: INFO: Retrying .... error trying to get Service mutability-test: Get "https://35.227.179.144/api/v1/namespaces/loadbalancers-6791/services/mutability-test": dial tcp 35.227.179.144:443: connect: connection refused Nov 26 05:32:09.063: INFO: Retrying .... error trying to get Service mutability-test: Get "https://35.227.179.144/api/v1/namespaces/loadbalancers-6791/services/mutability-test": dial tcp 35.227.179.144:443: connect: connection refused Nov 26 05:32:11.064: INFO: Retrying .... error trying to get Service mutability-test: Get "https://35.227.179.144/api/v1/namespaces/loadbalancers-6791/services/mutability-test": dial tcp 35.227.179.144:443: connect: connection refused Nov 26 05:32:13.063: INFO: Retrying .... error trying to get Service mutability-test: Get "https://35.227.179.144/api/v1/namespaces/loadbalancers-6791/services/mutability-test": dial tcp 35.227.179.144:443: connect: connection refused Nov 26 05:32:15.064: INFO: Retrying .... error trying to get Service mutability-test: Get "https://35.227.179.144/api/v1/namespaces/loadbalancers-6791/services/mutability-test": dial tcp 35.227.179.144:443: connect: connection refused Nov 26 05:32:17.063: INFO: Retrying .... error trying to get Service mutability-test: Get "https://35.227.179.144/api/v1/namespaces/loadbalancers-6791/services/mutability-test": dial tcp 35.227.179.144:443: connect: connection refused Nov 26 05:32:19.063: INFO: Retrying .... error trying to get Service mutability-test: Get "https://35.227.179.144/api/v1/namespaces/loadbalancers-6791/services/mutability-test": dial tcp 35.227.179.144:443: connect: connection refused ------------------------------ Progress Report for Ginkgo Process #6 Automatically polling progress: [sig-network] LoadBalancers should be able to change the type and ports of a UDP service [Slow] (Spec Runtime: 8m40.605s) test/e2e/network/loadbalancer.go:287 In [It] (Node Runtime: 8m40.043s) test/e2e/network/loadbalancer.go:287 At [By Step] waiting for the UDP service to have a load balancer (Step Runtime: 8m21.692s) test/e2e/network/loadbalancer.go:379 Spec Goroutine goroutine 1768 [select] k8s.io/kubernetes/vendor/k8s.io/apimachinery/pkg/util/wait.WaitForWithContext({0x7fe0bc8, 0xc0000820c8}, 0xc000e32cc0, 0x2fdb16a?) vendor/k8s.io/apimachinery/pkg/util/wait/wait.go:660 k8s.io/kubernetes/vendor/k8s.io/apimachinery/pkg/util/wait.poll({0x7fe0bc8, 0xc0000820c8}, 0x68?, 0x2fd9d05?, 0x20?) vendor/k8s.io/apimachinery/pkg/util/wait/wait.go:596 k8s.io/kubernetes/vendor/k8s.io/apimachinery/pkg/util/wait.PollImmediateWithContext({0x7fe0bc8, 0xc0000820c8}, 0xc0031e9b60?, 0xc003475bb8?, 0x262a967?) vendor/k8s.io/apimachinery/pkg/util/wait/wait.go:528 k8s.io/kubernetes/vendor/k8s.io/apimachinery/pkg/util/wait.PollImmediate(0xc000dd2130?, 0x7fa7740?, 0xc000202b80?) vendor/k8s.io/apimachinery/pkg/util/wait/wait.go:514 k8s.io/kubernetes/test/e2e/framework/service.(*TestJig).waitForCondition(0xc003b567d0, 0x4?, {0x7600fe2, 0x14}, 0x7895b68) test/e2e/framework/service/jig.go:631 k8s.io/kubernetes/test/e2e/framework/service.(*TestJig).WaitForLoadBalancer(0xc003b567d0, 0x33?) test/e2e/framework/service/jig.go:582 > k8s.io/kubernetes/test/e2e/network.glob..func19.4() test/e2e/network/loadbalancer.go:381 k8s.io/kubernetes/vendor/github.com/onsi/ginkgo/v2/internal.extractBodyFunction.func3({0x2d591ce, 0xc0056e0000}) vendor/github.com/onsi/ginkgo/v2/internal/node.go:449 k8s.io/kubernetes/vendor/github.com/onsi/ginkgo/v2/internal.(*Suite).runNode.func2() vendor/github.com/onsi/ginkgo/v2/internal/suite.go:750 k8s.io/kubernetes/vendor/github.com/onsi/ginkgo/v2/internal.(*Suite).runNode vendor/github.com/onsi/ginkgo/v2/internal/suite.go:738 ------------------------------ Nov 26 05:32:21.064: INFO: Retrying .... error trying to get Service mutability-test: Get "https://35.227.179.144/api/v1/namespaces/loadbalancers-6791/services/mutability-test": dial tcp 35.227.179.144:443: connect: connection refused Nov 26 05:32:23.063: INFO: Retrying .... error trying to get Service mutability-test: Get "https://35.227.179.144/api/v1/namespaces/loadbalancers-6791/services/mutability-test": dial tcp 35.227.179.144:443: connect: connection refused Nov 26 05:32:25.063: INFO: Retrying .... error trying to get Service mutability-test: Get "https://35.227.179.144/api/v1/namespaces/loadbalancers-6791/services/mutability-test": dial tcp 35.227.179.144:443: connect: connection refused Nov 26 05:32:27.063: INFO: Retrying .... error trying to get Service mutability-test: Get "https://35.227.179.144/api/v1/namespaces/loadbalancers-6791/services/mutability-test": dial tcp 35.227.179.144:443: connect: connection refused ------------------------------ Progress Report for Ginkgo Process #6 Automatically polling progress: [sig-network] LoadBalancers should be able to change the type and ports of a UDP service [Slow] (Spec Runtime: 9m0.607s) test/e2e/network/loadbalancer.go:287 In [It] (Node Runtime: 9m0.045s) test/e2e/network/loadbalancer.go:287 At [By Step] waiting for the UDP service to have a load balancer (Step Runtime: 8m41.694s) test/e2e/network/loadbalancer.go:379 Spec Goroutine goroutine 1768 [select] k8s.io/kubernetes/vendor/k8s.io/apimachinery/pkg/util/wait.WaitForWithContext({0x7fe0bc8, 0xc0000820c8}, 0xc000e32cc0, 0x2fdb16a?) vendor/k8s.io/apimachinery/pkg/util/wait/wait.go:660 k8s.io/kubernetes/vendor/k8s.io/apimachinery/pkg/util/wait.poll({0x7fe0bc8, 0xc0000820c8}, 0x68?, 0x2fd9d05?, 0x20?) vendor/k8s.io/apimachinery/pkg/util/wait/wait.go:596 k8s.io/kubernetes/vendor/k8s.io/apimachinery/pkg/util/wait.PollImmediateWithContext({0x7fe0bc8, 0xc0000820c8}, 0xc0031e9b60?, 0xc003475bb8?, 0x262a967?) vendor/k8s.io/apimachinery/pkg/util/wait/wait.go:528 k8s.io/kubernetes/vendor/k8s.io/apimachinery/pkg/util/wait.PollImmediate(0xc000dd2130?, 0x7fa7740?, 0xc000202b80?) vendor/k8s.io/apimachinery/pkg/util/wait/wait.go:514 k8s.io/kubernetes/test/e2e/framework/service.(*TestJig).waitForCondition(0xc003b567d0, 0x4?, {0x7600fe2, 0x14}, 0x7895b68) test/e2e/framework/service/jig.go:631 k8s.io/kubernetes/test/e2e/framework/service.(*TestJig).WaitForLoadBalancer(0xc003b567d0, 0x33?) test/e2e/framework/service/jig.go:582 > k8s.io/kubernetes/test/e2e/network.glob..func19.4() test/e2e/network/loadbalancer.go:381 k8s.io/kubernetes/vendor/github.com/onsi/ginkgo/v2/internal.extractBodyFunction.func3({0x2d591ce, 0xc0056e0000}) vendor/github.com/onsi/ginkgo/v2/internal/node.go:449 k8s.io/kubernetes/vendor/github.com/onsi/ginkgo/v2/internal.(*Suite).runNode.func2() vendor/github.com/onsi/ginkgo/v2/internal/suite.go:750 k8s.io/kubernetes/vendor/github.com/onsi/ginkgo/v2/internal.(*Suite).runNode vendor/github.com/onsi/ginkgo/v2/internal/suite.go:738 ------------------------------ ------------------------------ Progress Report for Ginkgo Process #6 Automatically polling progress: [sig-network] LoadBalancers should be able to change the type and ports of a UDP service [Slow] (Spec Runtime: 9m20.61s) test/e2e/network/loadbalancer.go:287 In [It] (Node Runtime: 9m20.047s) test/e2e/network/loadbalancer.go:287 At [By Step] waiting for the UDP service to have a load balancer (Step Runtime: 9m1.697s) test/e2e/network/loadbalancer.go:379 Spec Goroutine goroutine 1768 [select] k8s.io/kubernetes/vendor/k8s.io/apimachinery/pkg/util/wait.WaitForWithContext({0x7fe0bc8, 0xc0000820c8}, 0xc000e32cc0, 0x2fdb16a?) vendor/k8s.io/apimachinery/pkg/util/wait/wait.go:660 k8s.io/kubernetes/vendor/k8s.io/apimachinery/pkg/util/wait.poll({0x7fe0bc8, 0xc0000820c8}, 0x68?, 0x2fd9d05?, 0x20?) vendor/k8s.io/apimachinery/pkg/util/wait/wait.go:596 k8s.io/kubernetes/vendor/k8s.io/apimachinery/pkg/util/wait.PollImmediateWithContext({0x7fe0bc8, 0xc0000820c8}, 0xc0031e9b60?, 0xc003475bb8?, 0x262a967?) vendor/k8s.io/apimachinery/pkg/util/wait/wait.go:528 k8s.io/kubernetes/vendor/k8s.io/apimachinery/pkg/util/wait.PollImmediate(0xc000dd2130?, 0x7fa7740?, 0xc000202b80?) vendor/k8s.io/apimachinery/pkg/util/wait/wait.go:514 k8s.io/kubernetes/test/e2e/framework/service.(*TestJig).waitForCondition(0xc003b567d0, 0x4?, {0x7600fe2, 0x14}, 0x7895b68) test/e2e/framework/service/jig.go:631 k8s.io/kubernetes/test/e2e/framework/service.(*TestJig).WaitForLoadBalancer(0xc003b567d0, 0x33?) test/e2e/framework/service/jig.go:582 > k8s.io/kubernetes/test/e2e/network.glob..func19.4() test/e2e/network/loadbalancer.go:381 k8s.io/kubernetes/vendor/github.com/onsi/ginkgo/v2/internal.extractBodyFunction.func3({0x2d591ce, 0xc0056e0000}) vendor/github.com/onsi/ginkgo/v2/internal/node.go:449 k8s.io/kubernetes/vendor/github.com/onsi/ginkgo/v2/internal.(*Suite).runNode.func2() vendor/github.com/onsi/ginkgo/v2/internal/suite.go:750 k8s.io/kubernetes/vendor/github.com/onsi/ginkgo/v2/internal.(*Suite).runNode vendor/github.com/onsi/ginkgo/v2/internal/suite.go:738 ------------------------------ ------------------------------ Progress Report for Ginkgo Process #6 Automatically polling progress: [sig-network] LoadBalancers should be able to change the type and ports of a UDP service [Slow] (Spec Runtime: 9m40.612s) test/e2e/network/loadbalancer.go:287 In [It] (Node Runtime: 9m40.049s) test/e2e/network/loadbalancer.go:287 At [By Step] waiting for the UDP service to have a load balancer (Step Runtime: 9m21.698s) test/e2e/network/loadbalancer.go:379 Spec Goroutine goroutine 1768 [select] k8s.io/kubernetes/vendor/k8s.io/apimachinery/pkg/util/wait.WaitForWithContext({0x7fe0bc8, 0xc0000820c8}, 0xc000e32cc0, 0x2fdb16a?) vendor/k8s.io/apimachinery/pkg/util/wait/wait.go:660 k8s.io/kubernetes/vendor/k8s.io/apimachinery/pkg/util/wait.poll({0x7fe0bc8, 0xc0000820c8}, 0x68?, 0x2fd9d05?, 0x20?) vendor/k8s.io/apimachinery/pkg/util/wait/wait.go:596 k8s.io/kubernetes/vendor/k8s.io/apimachinery/pkg/util/wait.PollImmediateWithContext({0x7fe0bc8, 0xc0000820c8}, 0xc0031e9b60?, 0xc003475bb8?, 0x262a967?) vendor/k8s.io/apimachinery/pkg/util/wait/wait.go:528 k8s.io/kubernetes/vendor/k8s.io/apimachinery/pkg/util/wait.PollImmediate(0xc000dd2130?, 0x7fa7740?, 0xc000202b80?) vendor/k8s.io/apimachinery/pkg/util/wait/wait.go:514 k8s.io/kubernetes/test/e2e/framework/service.(*TestJig).waitForCondition(0xc003b567d0, 0x4?, {0x7600fe2, 0x14}, 0x7895b68) test/e2e/framework/service/jig.go:631 k8s.io/kubernetes/test/e2e/framework/service.(*TestJig).WaitForLoadBalancer(0xc003b567d0, 0x33?) test/e2e/framework/service/jig.go:582 > k8s.io/kubernetes/test/e2e/network.glob..func19.4() test/e2e/network/loadbalancer.go:381 k8s.io/kubernetes/vendor/github.com/onsi/ginkgo/v2/internal.extractBodyFunction.func3({0x2d591ce, 0xc0056e0000}) vendor/github.com/onsi/ginkgo/v2/internal/node.go:449 k8s.io/kubernetes/vendor/github.com/onsi/ginkgo/v2/internal.(*Suite).runNode.func2() vendor/github.com/onsi/ginkgo/v2/internal/suite.go:750 k8s.io/kubernetes/vendor/github.com/onsi/ginkgo/v2/internal.(*Suite).runNode vendor/github.com/onsi/ginkgo/v2/internal/suite.go:738 ------------------------------ ------------------------------ Progress Report for Ginkgo Process #6 Automatically polling progress: [sig-network] LoadBalancers should be able to change the type and ports of a UDP service [Slow] (Spec Runtime: 10m0.613s) test/e2e/network/loadbalancer.go:287 In [It] (Node Runtime: 10m0.051s) test/e2e/network/loadbalancer.go:287 At [By Step] waiting for the UDP service to have a load balancer (Step Runtime: 9m41.7s) test/e2e/network/loadbalancer.go:379 Spec Goroutine goroutine 1768 [select] k8s.io/kubernetes/vendor/k8s.io/apimachinery/pkg/util/wait.WaitForWithContext({0x7fe0bc8, 0xc0000820c8}, 0xc000e32cc0, 0x2fdb16a?) vendor/k8s.io/apimachinery/pkg/util/wait/wait.go:660 k8s.io/kubernetes/vendor/k8s.io/apimachinery/pkg/util/wait.poll({0x7fe0bc8, 0xc0000820c8}, 0x68?, 0x2fd9d05?, 0x20?) vendor/k8s.io/apimachinery/pkg/util/wait/wait.go:596 k8s.io/kubernetes/vendor/k8s.io/apimachinery/pkg/util/wait.PollImmediateWithContext({0x7fe0bc8, 0xc0000820c8}, 0xc0031e9b60?, 0xc003475bb8?, 0x262a967?) vendor/k8s.io/apimachinery/pkg/util/wait/wait.go:528 k8s.io/kubernetes/vendor/k8s.io/apimachinery/pkg/util/wait.PollImmediate(0xc000dd2130?, 0x7fa7740?, 0xc000202b80?) vendor/k8s.io/apimachinery/pkg/util/wait/wait.go:514 k8s.io/kubernetes/test/e2e/framework/service.(*TestJig).waitForCondition(0xc003b567d0, 0x4?, {0x7600fe2, 0x14}, 0x7895b68) test/e2e/framework/service/jig.go:631 k8s.io/kubernetes/test/e2e/framework/service.(*TestJig).WaitForLoadBalancer(0xc003b567d0, 0x33?) test/e2e/framework/service/jig.go:582 > k8s.io/kubernetes/test/e2e/network.glob..func19.4() test/e2e/network/loadbalancer.go:381 k8s.io/kubernetes/vendor/github.com/onsi/ginkgo/v2/internal.extractBodyFunction.func3({0x2d591ce, 0xc0056e0000}) vendor/github.com/onsi/ginkgo/v2/internal/node.go:449 k8s.io/kubernetes/vendor/github.com/onsi/ginkgo/v2/internal.(*Suite).runNode.func2() vendor/github.com/onsi/ginkgo/v2/internal/suite.go:750 k8s.io/kubernetes/vendor/github.com/onsi/ginkgo/v2/internal.(*Suite).runNode vendor/github.com/onsi/ginkgo/v2/internal/suite.go:738 ------------------------------ ------------------------------ Progress Report for Ginkgo Process #6 Automatically polling progress: [sig-network] LoadBalancers should be able to change the type and ports of a UDP service [Slow] (Spec Runtime: 10m20.616s) test/e2e/network/loadbalancer.go:287 In [It] (Node Runtime: 10m20.054s) test/e2e/network/loadbalancer.go:287 At [By Step] waiting for the UDP service to have a load balancer (Step Runtime: 10m1.703s) test/e2e/network/loadbalancer.go:379 Spec Goroutine goroutine 1768 [select] k8s.io/kubernetes/vendor/k8s.io/apimachinery/pkg/util/wait.WaitForWithContext({0x7fe0bc8, 0xc0000820c8}, 0xc000e32cc0, 0x2fdb16a?) vendor/k8s.io/apimachinery/pkg/util/wait/wait.go:660 k8s.io/kubernetes/vendor/k8s.io/apimachinery/pkg/util/wait.poll({0x7fe0bc8, 0xc0000820c8}, 0x68?, 0x2fd9d05?, 0x20?) vendor/k8s.io/apimachinery/pkg/util/wait/wait.go:596 k8s.io/kubernetes/vendor/k8s.io/apimachinery/pkg/util/wait.PollImmediateWithContext({0x7fe0bc8, 0xc0000820c8}, 0xc0031e9b60?, 0xc003475bb8?, 0x262a967?) vendor/k8s.io/apimachinery/pkg/util/wait/wait.go:528 k8s.io/kubernetes/vendor/k8s.io/apimachinery/pkg/util/wait.PollImmediate(0xc000dd2130?, 0x7fa7740?, 0xc000202b80?) vendor/k8s.io/apimachinery/pkg/util/wait/wait.go:514 k8s.io/kubernetes/test/e2e/framework/service.(*TestJig).waitForCondition(0xc003b567d0, 0x4?, {0x7600fe2, 0x14}, 0x7895b68) test/e2e/framework/service/jig.go:631 k8s.io/kubernetes/test/e2e/framework/service.(*TestJig).WaitForLoadBalancer(0xc003b567d0, 0x33?) test/e2e/framework/service/jig.go:582 > k8s.io/kubernetes/test/e2e/network.glob..func19.4() test/e2e/network/loadbalancer.go:381 k8s.io/kubernetes/vendor/github.com/onsi/ginkgo/v2/internal.extractBodyFunction.func3({0x2d591ce, 0xc0056e0000}) vendor/github.com/onsi/ginkgo/v2/internal/node.go:449 k8s.io/kubernetes/vendor/github.com/onsi/ginkgo/v2/internal.(*Suite).runNode.func2() vendor/github.com/onsi/ginkgo/v2/internal/suite.go:750 k8s.io/kubernetes/vendor/github.com/onsi/ginkgo/v2/internal.(*Suite).runNode vendor/github.com/onsi/ginkgo/v2/internal/suite.go:738 ------------------------------ ------------------------------ Progress Report for Ginkgo Process #6 Automatically polling progress: [sig-network] LoadBalancers should be able to change the type and ports of a UDP service [Slow] (Spec Runtime: 10m40.62s) test/e2e/network/loadbalancer.go:287 In [It] (Node Runtime: 10m40.057s) test/e2e/network/loadbalancer.go:287 At [By Step] waiting for the UDP service to have a load balancer (Step Runtime: 10m21.706s) test/e2e/network/loadbalancer.go:379 Spec Goroutine goroutine 1768 [select] k8s.io/kubernetes/vendor/k8s.io/apimachinery/pkg/util/wait.WaitForWithContext({0x7fe0bc8, 0xc0000820c8}, 0xc000e32cc0, 0x2fdb16a?) vendor/k8s.io/apimachinery/pkg/util/wait/wait.go:660 k8s.io/kubernetes/vendor/k8s.io/apimachinery/pkg/util/wait.poll({0x7fe0bc8, 0xc0000820c8}, 0x68?, 0x2fd9d05?, 0x20?) vendor/k8s.io/apimachinery/pkg/util/wait/wait.go:596 k8s.io/kubernetes/vendor/k8s.io/apimachinery/pkg/util/wait.PollImmediateWithContext({0x7fe0bc8, 0xc0000820c8}, 0xc0031e9b60?, 0xc003475bb8?, 0x262a967?) vendor/k8s.io/apimachinery/pkg/util/wait/wait.go:528 k8s.io/kubernetes/vendor/k8s.io/apimachinery/pkg/util/wait.PollImmediate(0xc000dd2130?, 0x7fa7740?, 0xc000202b80?) vendor/k8s.io/apimachinery/pkg/util/wait/wait.go:514 k8s.io/kubernetes/test/e2e/framework/service.(*TestJig).waitForCondition(0xc003b567d0, 0x4?, {0x7600fe2, 0x14}, 0x7895b68) test/e2e/framework/service/jig.go:631 k8s.io/kubernetes/test/e2e/framework/service.(*TestJig).WaitForLoadBalancer(0xc003b567d0, 0x33?) test/e2e/framework/service/jig.go:582 > k8s.io/kubernetes/test/e2e/network.glob..func19.4() test/e2e/network/loadbalancer.go:381 k8s.io/kubernetes/vendor/github.com/onsi/ginkgo/v2/internal.extractBodyFunction.func3({0x2d591ce, 0xc0056e0000}) vendor/github.com/onsi/ginkgo/v2/internal/node.go:449 k8s.io/kubernetes/vendor/github.com/onsi/ginkgo/v2/internal.(*Suite).runNode.func2() vendor/github.com/onsi/ginkgo/v2/internal/suite.go:750 k8s.io/kubernetes/vendor/github.com/onsi/ginkgo/v2/internal.(*Suite).runNode vendor/github.com/onsi/ginkgo/v2/internal/suite.go:738 ------------------------------ ------------------------------ Progress Report for Ginkgo Process #6 Automatically polling progress: [sig-network] LoadBalancers should be able to change the type and ports of a UDP service [Slow] (Spec Runtime: 11m0.622s) test/e2e/network/loadbalancer.go:287 In [It] (Node Runtime: 11m0.06s) test/e2e/network/loadbalancer.go:287 At [By Step] waiting for the UDP service to have a load balancer (Step Runtime: 10m41.709s) test/e2e/network/loadbalancer.go:379 Spec Goroutine goroutine 1768 [select] k8s.io/kubernetes/vendor/k8s.io/apimachinery/pkg/util/wait.WaitForWithContext({0x7fe0bc8, 0xc0000820c8}, 0xc000e32cc0, 0x2fdb16a?) vendor/k8s.io/apimachinery/pkg/util/wait/wait.go:660 k8s.io/kubernetes/vendor/k8s.io/apimachinery/pkg/util/wait.poll({0x7fe0bc8, 0xc0000820c8}, 0x68?, 0x2fd9d05?, 0x20?) vendor/k8s.io/apimachinery/pkg/util/wait/wait.go:596 k8s.io/kubernetes/vendor/k8s.io/apimachinery/pkg/util/wait.PollImmediateWithContext({0x7fe0bc8, 0xc0000820c8}, 0xc0031e9b60?, 0xc003475bb8?, 0x262a967?) vendor/k8s.io/apimachinery/pkg/util/wait/wait.go:528 k8s.io/kubernetes/vendor/k8s.io/apimachinery/pkg/util/wait.PollImmediate(0xc000dd2130?, 0x7fa7740?, 0xc000202b80?) vendor/k8s.io/apimachinery/pkg/util/wait/wait.go:514 k8s.io/kubernetes/test/e2e/framework/service.(*TestJig).waitForCondition(0xc003b567d0, 0x4?, {0x7600fe2, 0x14}, 0x7895b68) test/e2e/framework/service/jig.go:631 k8s.io/kubernetes/test/e2e/framework/service.(*TestJig).WaitForLoadBalancer(0xc003b567d0, 0x33?) test/e2e/framework/service/jig.go:582 > k8s.io/kubernetes/test/e2e/network.glob..func19.4() test/e2e/network/loadbalancer.go:381 k8s.io/kubernetes/vendor/github.com/onsi/ginkgo/v2/internal.extractBodyFunction.func3({0x2d591ce, 0xc0056e0000}) vendor/github.com/onsi/ginkgo/v2/internal/node.go:449 k8s.io/kubernetes/vendor/github.com/onsi/ginkgo/v2/internal.(*Suite).runNode.func2() vendor/github.com/onsi/ginkgo/v2/internal/suite.go:750 k8s.io/kubernetes/vendor/github.com/onsi/ginkgo/v2/internal.(*Suite).runNode vendor/github.com/onsi/ginkgo/v2/internal/suite.go:738 ------------------------------ ------------------------------ Progress Report for Ginkgo Process #6 Automatically polling progress: [sig-network] LoadBalancers should be able to change the type and ports of a UDP service [Slow] (Spec Runtime: 11m20.624s) test/e2e/network/loadbalancer.go:287 In [It] (Node Runtime: 11m20.061s) test/e2e/network/loadbalancer.go:287 At [By Step] waiting for the UDP service to have a load balancer (Step Runtime: 11m1.71s) test/e2e/network/loadbalancer.go:379 Spec Goroutine goroutine 1768 [select] k8s.io/kubernetes/vendor/k8s.io/apimachinery/pkg/util/wait.WaitForWithContext({0x7fe0bc8, 0xc0000820c8}, 0xc000e32cc0, 0x2fdb16a?) vendor/k8s.io/apimachinery/pkg/util/wait/wait.go:660 k8s.io/kubernetes/vendor/k8s.io/apimachinery/pkg/util/wait.poll({0x7fe0bc8, 0xc0000820c8}, 0x68?, 0x2fd9d05?, 0x20?) vendor/k8s.io/apimachinery/pkg/util/wait/wait.go:596 k8s.io/kubernetes/vendor/k8s.io/apimachinery/pkg/util/wait.PollImmediateWithContext({0x7fe0bc8, 0xc0000820c8}, 0xc0031e9b60?, 0xc003475bb8?, 0x262a967?) vendor/k8s.io/apimachinery/pkg/util/wait/wait.go:528 k8s.io/kubernetes/vendor/k8s.io/apimachinery/pkg/util/wait.PollImmediate(0xc000dd2130?, 0x7fa7740?, 0xc000202b80?) vendor/k8s.io/apimachinery/pkg/util/wait/wait.go:514 k8s.io/kubernetes/test/e2e/framework/service.(*TestJig).waitForCondition(0xc003b567d0, 0x4?, {0x7600fe2, 0x14}, 0x7895b68) test/e2e/framework/service/jig.go:631 k8s.io/kubernetes/test/e2e/framework/service.(*TestJig).WaitForLoadBalancer(0xc003b567d0, 0x33?) test/e2e/framework/service/jig.go:582 > k8s.io/kubernetes/test/e2e/network.glob..func19.4() test/e2e/network/loadbalancer.go:381 k8s.io/kubernetes/vendor/github.com/onsi/ginkgo/v2/internal.extractBodyFunction.func3({0x2d591ce, 0xc0056e0000}) vendor/github.com/onsi/ginkgo/v2/internal/node.go:449 k8s.io/kubernetes/vendor/github.com/onsi/ginkgo/v2/internal.(*Suite).runNode.func2() vendor/github.com/onsi/ginkgo/v2/internal/suite.go:750 k8s.io/kubernetes/vendor/github.com/onsi/ginkgo/v2/internal.(*Suite).runNode vendor/github.com/onsi/ginkgo/v2/internal/suite.go:738 ------------------------------ ------------------------------ Progress Report for Ginkgo Process #6 Automatically polling progress: [sig-network] LoadBalancers should be able to change the type and ports of a UDP service [Slow] (Spec Runtime: 11m40.626s) test/e2e/network/loadbalancer.go:287 In [It] (Node Runtime: 11m40.063s) test/e2e/network/loadbalancer.go:287 At [By Step] waiting for the UDP service to have a load balancer (Step Runtime: 11m21.713s) test/e2e/network/loadbalancer.go:379 Spec Goroutine goroutine 1768 [select] k8s.io/kubernetes/vendor/k8s.io/apimachinery/pkg/util/wait.WaitForWithContext({0x7fe0bc8, 0xc0000820c8}, 0xc000e32cc0, 0x2fdb16a?) vendor/k8s.io/apimachinery/pkg/util/wait/wait.go:660 k8s.io/kubernetes/vendor/k8s.io/apimachinery/pkg/util/wait.poll({0x7fe0bc8, 0xc0000820c8}, 0x68?, 0x2fd9d05?, 0x20?) vendor/k8s.io/apimachinery/pkg/util/wait/wait.go:596 k8s.io/kubernetes/vendor/k8s.io/apimachinery/pkg/util/wait.PollImmediateWithContext({0x7fe0bc8, 0xc0000820c8}, 0xc0031e9b60?, 0xc003475bb8?, 0x262a967?) vendor/k8s.io/apimachinery/pkg/util/wait/wait.go:528 k8s.io/kubernetes/vendor/k8s.io/apimachinery/pkg/util/wait.PollImmediate(0xc000dd2130?, 0x7fa7740?, 0xc000202b80?) vendor/k8s.io/apimachinery/pkg/util/wait/wait.go:514 k8s.io/kubernetes/test/e2e/framework/service.(*TestJig).waitForCondition(0xc003b567d0, 0x4?, {0x7600fe2, 0x14}, 0x7895b68) test/e2e/framework/service/jig.go:631 k8s.io/kubernetes/test/e2e/framework/service.(*TestJig).WaitForLoadBalancer(0xc003b567d0, 0x33?) test/e2e/framework/service/jig.go:582 > k8s.io/kubernetes/test/e2e/network.glob..func19.4() test/e2e/network/loadbalancer.go:381 k8s.io/kubernetes/vendor/github.com/onsi/ginkgo/v2/internal.extractBodyFunction.func3({0x2d591ce, 0xc0056e0000}) vendor/github.com/onsi/ginkgo/v2/internal/node.go:449 k8s.io/kubernetes/vendor/github.com/onsi/ginkgo/v2/internal.(*Suite).runNode.func2() vendor/github.com/onsi/ginkgo/v2/internal/suite.go:750 k8s.io/kubernetes/vendor/github.com/onsi/ginkgo/v2/internal.(*Suite).runNode vendor/github.com/onsi/ginkgo/v2/internal/suite.go:738 ------------------------------ ------------------------------ Progress Report for Ginkgo Process #6 Automatically polling progress: [sig-network] LoadBalancers should be able to change the type and ports of a UDP service [Slow] (Spec Runtime: 12m0.628s) test/e2e/network/loadbalancer.go:287 In [It] (Node Runtime: 12m0.065s) test/e2e/network/loadbalancer.go:287 At [By Step] waiting for the UDP service to have a load balancer (Step Runtime: 11m41.714s) test/e2e/network/loadbalancer.go:379 Spec Goroutine goroutine 1768 [select] k8s.io/kubernetes/vendor/k8s.io/apimachinery/pkg/util/wait.WaitForWithContext({0x7fe0bc8, 0xc0000820c8}, 0xc000e32cc0, 0x2fdb16a?) vendor/k8s.io/apimachinery/pkg/util/wait/wait.go:660 k8s.io/kubernetes/vendor/k8s.io/apimachinery/pkg/util/wait.poll({0x7fe0bc8, 0xc0000820c8}, 0x68?, 0x2fd9d05?, 0x20?) vendor/k8s.io/apimachinery/pkg/util/wait/wait.go:596 k8s.io/kubernetes/vendor/k8s.io/apimachinery/pkg/util/wait.PollImmediateWithContext({0x7fe0bc8, 0xc0000820c8}, 0xc0031e9b60?, 0xc003475bb8?, 0x262a967?) vendor/k8s.io/apimachinery/pkg/util/wait/wait.go:528 k8s.io/kubernetes/vendor/k8s.io/apimachinery/pkg/util/wait.PollImmediate(0xc000dd2130?, 0x7fa7740?, 0xc000202b80?) vendor/k8s.io/apimachinery/pkg/util/wait/wait.go:514 k8s.io/kubernetes/test/e2e/framework/service.(*TestJig).waitForCondition(0xc003b567d0, 0x4?, {0x7600fe2, 0x14}, 0x7895b68) test/e2e/framework/service/jig.go:631 k8s.io/kubernetes/test/e2e/framework/service.(*TestJig).WaitForLoadBalancer(0xc003b567d0, 0x33?) test/e2e/framework/service/jig.go:582 > k8s.io/kubernetes/test/e2e/network.glob..func19.4() test/e2e/network/loadbalancer.go:381 k8s.io/kubernetes/vendor/github.com/onsi/ginkgo/v2/internal.extractBodyFunction.func3({0x2d591ce, 0xc0056e0000}) vendor/github.com/onsi/ginkgo/v2/internal/node.go:449 k8s.io/kubernetes/vendor/github.com/onsi/ginkgo/v2/internal.(*Suite).runNode.func2() vendor/github.com/onsi/ginkgo/v2/internal/suite.go:750 k8s.io/kubernetes/vendor/github.com/onsi/ginkgo/v2/internal.(*Suite).runNode vendor/github.com/onsi/ginkgo/v2/internal/suite.go:738 ------------------------------ ------------------------------ Progress Report for Ginkgo Process #6 Automatically polling progress: [sig-network] LoadBalancers should be able to change the type and ports of a UDP service [Slow] (Spec Runtime: 12m20.629s) test/e2e/network/loadbalancer.go:287 In [It] (Node Runtime: 12m20.067s) test/e2e/network/loadbalancer.go:287 At [By Step] waiting for the UDP service to have a load balancer (Step Runtime: 12m1.716s) test/e2e/network/loadbalancer.go:379 Spec Goroutine goroutine 1768 [select] k8s.io/kubernetes/vendor/k8s.io/apimachinery/pkg/util/wait.WaitForWithContext({0x7fe0bc8, 0xc0000820c8}, 0xc000e32cc0, 0x2fdb16a?) vendor/k8s.io/apimachinery/pkg/util/wait/wait.go:660 k8s.io/kubernetes/vendor/k8s.io/apimachinery/pkg/util/wait.poll({0x7fe0bc8, 0xc0000820c8}, 0x68?, 0x2fd9d05?, 0x20?) vendor/k8s.io/apimachinery/pkg/util/wait/wait.go:596 k8s.io/kubernetes/vendor/k8s.io/apimachinery/pkg/util/wait.PollImmediateWithContext({0x7fe0bc8, 0xc0000820c8}, 0xc0031e9b60?, 0xc003475bb8?, 0x262a967?) vendor/k8s.io/apimachinery/pkg/util/wait/wait.go:528 k8s.io/kubernetes/vendor/k8s.io/apimachinery/pkg/util/wait.PollImmediate(0xc000dd2130?, 0x7fa7740?, 0xc000202b80?) vendor/k8s.io/apimachinery/pkg/util/wait/wait.go:514 k8s.io/kubernetes/test/e2e/framework/service.(*TestJig).waitForCondition(0xc003b567d0, 0x4?, {0x7600fe2, 0x14}, 0x7895b68) test/e2e/framework/service/jig.go:631 k8s.io/kubernetes/test/e2e/framework/service.(*TestJig).WaitForLoadBalancer(0xc003b567d0, 0x33?) test/e2e/framework/service/jig.go:582 > k8s.io/kubernetes/test/e2e/network.glob..func19.4() test/e2e/network/loadbalancer.go:381 k8s.io/kubernetes/vendor/github.com/onsi/ginkgo/v2/internal.extractBodyFunction.func3({0x2d591ce, 0xc0056e0000}) vendor/github.com/onsi/ginkgo/v2/internal/node.go:449 k8s.io/kubernetes/vendor/github.com/onsi/ginkgo/v2/internal.(*Suite).runNode.func2() vendor/github.com/onsi/ginkgo/v2/internal/suite.go:750 k8s.io/kubernetes/vendor/github.com/onsi/ginkgo/v2/internal.(*Suite).runNode vendor/github.com/onsi/ginkgo/v2/internal/suite.go:738 ------------------------------ ------------------------------ Progress Report for Ginkgo Process #6 Automatically polling progress: [sig-network] LoadBalancers should be able to change the type and ports of a UDP service [Slow] (Spec Runtime: 12m40.632s) test/e2e/network/loadbalancer.go:287 In [It] (Node Runtime: 12m40.069s) test/e2e/network/loadbalancer.go:287 At [By Step] waiting for the UDP service to have a load balancer (Step Runtime: 12m21.719s) test/e2e/network/loadbalancer.go:379 Spec Goroutine goroutine 1768 [select] k8s.io/kubernetes/vendor/k8s.io/apimachinery/pkg/util/wait.WaitForWithContext({0x7fe0bc8, 0xc0000820c8}, 0xc000e32cc0, 0x2fdb16a?) vendor/k8s.io/apimachinery/pkg/util/wait/wait.go:660 k8s.io/kubernetes/vendor/k8s.io/apimachinery/pkg/util/wait.poll({0x7fe0bc8, 0xc0000820c8}, 0x68?, 0x2fd9d05?, 0x20?) vendor/k8s.io/apimachinery/pkg/util/wait/wait.go:596 k8s.io/kubernetes/vendor/k8s.io/apimachinery/pkg/util/wait.PollImmediateWithContext({0x7fe0bc8, 0xc0000820c8}, 0xc0031e9b60?, 0xc003475bb8?, 0x262a967?) vendor/k8s.io/apimachinery/pkg/util/wait/wait.go:528 k8s.io/kubernetes/vendor/k8s.io/apimachinery/pkg/util/wait.PollImmediate(0xc000dd2130?, 0x7fa7740?, 0xc000202b80?) vendor/k8s.io/apimachinery/pkg/util/wait/wait.go:514 k8s.io/kubernetes/test/e2e/framework/service.(*TestJig).waitForCondition(0xc003b567d0, 0x4?, {0x7600fe2, 0x14}, 0x7895b68) test/e2e/framework/service/jig.go:631 k8s.io/kubernetes/test/e2e/framework/service.(*TestJig).WaitForLoadBalancer(0xc003b567d0, 0x33?) test/e2e/framework/service/jig.go:582 > k8s.io/kubernetes/test/e2e/network.glob..func19.4() test/e2e/network/loadbalancer.go:381 k8s.io/kubernetes/vendor/github.com/onsi/ginkgo/v2/internal.extractBodyFunction.func3({0x2d591ce, 0xc0056e0000}) vendor/github.com/onsi/ginkgo/v2/internal/node.go:449 k8s.io/kubernetes/vendor/github.com/onsi/ginkgo/v2/internal.(*Suite).runNode.func2() vendor/github.com/onsi/ginkgo/v2/internal/suite.go:750 k8s.io/kubernetes/vendor/github.com/onsi/ginkgo/v2/internal.(*Suite).runNode vendor/github.com/onsi/ginkgo/v2/internal/suite.go:738 ------------------------------ ------------------------------ Progress Report for Ginkgo Process #6 Automatically polling progress: [sig-network] LoadBalancers should be able to change the type and ports of a UDP service [Slow] (Spec Runtime: 13m0.634s) test/e2e/network/loadbalancer.go:287 In [It] (Node Runtime: 13m0.072s) test/e2e/network/loadbalancer.go:287 At [By Step] waiting for the UDP service to have a load balancer (Step Runtime: 12m41.721s) test/e2e/network/loadbalancer.go:379 Spec Goroutine goroutine 1768 [select] k8s.io/kubernetes/vendor/k8s.io/apimachinery/pkg/util/wait.WaitForWithContext({0x7fe0bc8, 0xc0000820c8}, 0xc000e32cc0, 0x2fdb16a?) vendor/k8s.io/apimachinery/pkg/util/wait/wait.go:660 k8s.io/kubernetes/vendor/k8s.io/apimachinery/pkg/util/wait.poll({0x7fe0bc8, 0xc0000820c8}, 0x68?, 0x2fd9d05?, 0x20?) vendor/k8s.io/apimachinery/pkg/util/wait/wait.go:596 k8s.io/kubernetes/vendor/k8s.io/apimachinery/pkg/util/wait.PollImmediateWithContext({0x7fe0bc8, 0xc0000820c8}, 0xc0031e9b60?, 0xc003475bb8?, 0x262a967?) vendor/k8s.io/apimachinery/pkg/util/wait/wait.go:528 k8s.io/kubernetes/vendor/k8s.io/apimachinery/pkg/util/wait.PollImmediate(0xc000dd2130?, 0x7fa7740?, 0xc000202b80?) vendor/k8s.io/apimachinery/pkg/util/wait/wait.go:514 k8s.io/kubernetes/test/e2e/framework/service.(*TestJig).waitForCondition(0xc003b567d0, 0x4?, {0x7600fe2, 0x14}, 0x7895b68) test/e2e/framework/service/jig.go:631 k8s.io/kubernetes/test/e2e/framework/service.(*TestJig).WaitForLoadBalancer(0xc003b567d0, 0x33?) test/e2e/framework/service/jig.go:582 > k8s.io/kubernetes/test/e2e/network.glob..func19.4() test/e2e/network/loadbalancer.go:381 k8s.io/kubernetes/vendor/github.com/onsi/ginkgo/v2/internal.extractBodyFunction.func3({0x2d591ce, 0xc0056e0000}) vendor/github.com/onsi/ginkgo/v2/internal/node.go:449 k8s.io/kubernetes/vendor/github.com/onsi/ginkgo/v2/internal.(*Suite).runNode.func2() vendor/github.com/onsi/ginkgo/v2/internal/suite.go:750 k8s.io/kubernetes/vendor/github.com/onsi/ginkgo/v2/internal.(*Suite).runNode vendor/github.com/onsi/ginkgo/v2/internal/suite.go:738 ------------------------------ ------------------------------ Progress Report for Ginkgo Process #6 Automatically polling progress: [sig-network] LoadBalancers should be able to change the type and ports of a UDP service [Slow] (Spec Runtime: 13m20.636s) test/e2e/network/loadbalancer.go:287 In [It] (Node Runtime: 13m20.073s) test/e2e/network/loadbalancer.go:287 At [By Step] waiting for the UDP service to have a load balancer (Step Runtime: 13m1.723s) test/e2e/network/loadbalancer.go:379 Spec Goroutine goroutine 1768 [select] k8s.io/kubernetes/vendor/k8s.io/apimachinery/pkg/util/wait.WaitForWithContext({0x7fe0bc8, 0xc0000820c8}, 0xc000e32cc0, 0x2fdb16a?) vendor/k8s.io/apimachinery/pkg/util/wait/wait.go:660 k8s.io/kubernetes/vendor/k8s.io/apimachinery/pkg/util/wait.poll({0x7fe0bc8, 0xc0000820c8}, 0x68?, 0x2fd9d05?, 0x20?) vendor/k8s.io/apimachinery/pkg/util/wait/wait.go:596 k8s.io/kubernetes/vendor/k8s.io/apimachinery/pkg/util/wait.PollImmediateWithContext({0x7fe0bc8, 0xc0000820c8}, 0xc0031e9b60?, 0xc003475bb8?, 0x262a967?) vendor/k8s.io/apimachinery/pkg/util/wait/wait.go:528 k8s.io/kubernetes/vendor/k8s.io/apimachinery/pkg/util/wait.PollImmediate(0xc000dd2130?, 0x7fa7740?, 0xc000202b80?) vendor/k8s.io/apimachinery/pkg/util/wait/wait.go:514 k8s.io/kubernetes/test/e2e/framework/service.(*TestJig).waitForCondition(0xc003b567d0, 0x4?, {0x7600fe2, 0x14}, 0x7895b68) test/e2e/framework/service/jig.go:631 k8s.io/kubernetes/test/e2e/framework/service.(*TestJig).WaitForLoadBalancer(0xc003b567d0, 0x33?) test/e2e/framework/service/jig.go:582 > k8s.io/kubernetes/test/e2e/network.glob..func19.4() test/e2e/network/loadbalancer.go:381 k8s.io/kubernetes/vendor/github.com/onsi/ginkgo/v2/internal.extractBodyFunction.func3({0x2d591ce, 0xc0056e0000}) vendor/github.com/onsi/ginkgo/v2/internal/node.go:449 k8s.io/kubernetes/vendor/github.com/onsi/ginkgo/v2/internal.(*Suite).runNode.func2() vendor/github.com/onsi/ginkgo/v2/internal/suite.go:750 k8s.io/kubernetes/vendor/github.com/onsi/ginkgo/v2/internal.(*Suite).runNode vendor/github.com/onsi/ginkgo/v2/internal/suite.go:738 ------------------------------ ------------------------------ Progress Report for Ginkgo Process #6 Automatically polling progress: [sig-network] LoadBalancers should be able to change the type and ports of a UDP service [Slow] (Spec Runtime: 13m40.638s) test/e2e/network/loadbalancer.go:287 In [It] (Node Runtime: 13m40.076s) test/e2e/network/loadbalancer.go:287 At [By Step] waiting for the UDP service to have a load balancer (Step Runtime: 13m21.725s) test/e2e/network/loadbalancer.go:379 Spec Goroutine goroutine 1768 [select] k8s.io/kubernetes/vendor/k8s.io/apimachinery/pkg/util/wait.WaitForWithContext({0x7fe0bc8, 0xc0000820c8}, 0xc000e32cc0, 0x2fdb16a?) vendor/k8s.io/apimachinery/pkg/util/wait/wait.go:660 k8s.io/kubernetes/vendor/k8s.io/apimachinery/pkg/util/wait.poll({0x7fe0bc8, 0xc0000820c8}, 0x68?, 0x2fd9d05?, 0x20?) vendor/k8s.io/apimachinery/pkg/util/wait/wait.go:596 k8s.io/kubernetes/vendor/k8s.io/apimachinery/pkg/util/wait.PollImmediateWithContext({0x7fe0bc8, 0xc0000820c8}, 0xc0031e9b60?, 0xc003475bb8?, 0x262a967?) vendor/k8s.io/apimachinery/pkg/util/wait/wait.go:528 k8s.io/kubernetes/vendor/k8s.io/apimachinery/pkg/util/wait.PollImmediate(0xc000dd2130?, 0x7fa7740?, 0xc000202b80?) vendor/k8s.io/apimachinery/pkg/util/wait/wait.go:514 k8s.io/kubernetes/test/e2e/framework/service.(*TestJig).waitForCondition(0xc003b567d0, 0x4?, {0x7600fe2, 0x14}, 0x7895b68) test/e2e/framework/service/jig.go:631 k8s.io/kubernetes/test/e2e/framework/service.(*TestJig).WaitForLoadBalancer(0xc003b567d0, 0x33?) test/e2e/framework/service/jig.go:582 > k8s.io/kubernetes/test/e2e/network.glob..func19.4() test/e2e/network/loadbalancer.go:381 k8s.io/kubernetes/vendor/github.com/onsi/ginkgo/v2/internal.extractBodyFunction.func3({0x2d591ce, 0xc0056e0000}) vendor/github.com/onsi/ginkgo/v2/internal/node.go:449 k8s.io/kubernetes/vendor/github.com/onsi/ginkgo/v2/internal.(*Suite).runNode.func2() vendor/github.com/onsi/ginkgo/v2/internal/suite.go:750 k8s.io/kubernetes/vendor/github.com/onsi/ginkgo/v2/internal.(*Suite).runNode vendor/github.com/onsi/ginkgo/v2/internal/suite.go:738 ------------------------------ ------------------------------ Progress Report for Ginkgo Process #6 Automatically polling progress: [sig-network] LoadBalancers should be able to change the type and ports of a UDP service [Slow] (Spec Runtime: 14m0.641s) test/e2e/network/loadbalancer.go:287 In [It] (Node Runtime: 14m0.078s) test/e2e/network/loadbalancer.go:287 At [By Step] waiting for the UDP service to have a load balancer (Step Runtime: 13m41.727s) test/e2e/network/loadbalancer.go:379 Spec Goroutine goroutine 1768 [select] k8s.io/kubernetes/vendor/k8s.io/apimachinery/pkg/util/wait.WaitForWithContext({0x7fe0bc8, 0xc0000820c8}, 0xc000e32cc0, 0x2fdb16a?) vendor/k8s.io/apimachinery/pkg/util/wait/wait.go:660 k8s.io/kubernetes/vendor/k8s.io/apimachinery/pkg/util/wait.poll({0x7fe0bc8, 0xc0000820c8}, 0x68?, 0x2fd9d05?, 0x20?) vendor/k8s.io/apimachinery/pkg/util/wait/wait.go:596 k8s.io/kubernetes/vendor/k8s.io/apimachinery/pkg/util/wait.PollImmediateWithContext({0x7fe0bc8, 0xc0000820c8}, 0xc0031e9b60?, 0xc003475bb8?, 0x262a967?) vendor/k8s.io/apimachinery/pkg/util/wait/wait.go:528 k8s.io/kubernetes/vendor/k8s.io/apimachinery/pkg/util/wait.PollImmediate(0xc000dd2130?, 0x7fa7740?, 0xc000202b80?) vendor/k8s.io/apimachinery/pkg/util/wait/wait.go:514 k8s.io/kubernetes/test/e2e/framework/service.(*TestJig).waitForCondition(0xc003b567d0, 0x4?, {0x7600fe2, 0x14}, 0x7895b68) test/e2e/framework/service/jig.go:631 k8s.io/kubernetes/test/e2e/framework/service.(*TestJig).WaitForLoadBalancer(0xc003b567d0, 0x33?) test/e2e/framework/service/jig.go:582 > k8s.io/kubernetes/test/e2e/network.glob..func19.4() test/e2e/network/loadbalancer.go:381 k8s.io/kubernetes/vendor/github.com/onsi/ginkgo/v2/internal.extractBodyFunction.func3({0x2d591ce, 0xc0056e0000}) vendor/github.com/onsi/ginkgo/v2/internal/node.go:449 k8s.io/kubernetes/vendor/github.com/onsi/ginkgo/v2/internal.(*Suite).runNode.func2() vendor/github.com/onsi/ginkgo/v2/internal/suite.go:750 k8s.io/kubernetes/vendor/github.com/onsi/ginkgo/v2/internal.(*Suite).runNode vendor/github.com/onsi/ginkgo/v2/internal/suite.go:738 ------------------------------ ------------------------------ Progress Report for Ginkgo Process #6 Automatically polling progress: [sig-network] LoadBalancers should be able to change the type and ports of a UDP service [Slow] (Spec Runtime: 14m20.643s) test/e2e/network/loadbalancer.go:287 In [It] (Node Runtime: 14m20.08s) test/e2e/network/loadbalancer.go:287 At [By Step] waiting for the UDP service to have a load balancer (Step Runtime: 14m1.73s) test/e2e/network/loadbalancer.go:379 Spec Goroutine goroutine 1768 [select] k8s.io/kubernetes/vendor/k8s.io/apimachinery/pkg/util/wait.WaitForWithContext({0x7fe0bc8, 0xc0000820c8}, 0xc000e32cc0, 0x2fdb16a?) vendor/k8s.io/apimachinery/pkg/util/wait/wait.go:660 k8s.io/kubernetes/vendor/k8s.io/apimachinery/pkg/util/wait.poll({0x7fe0bc8, 0xc0000820c8}, 0x68?, 0x2fd9d05?, 0x20?) vendor/k8s.io/apimachinery/pkg/util/wait/wait.go:596 k8s.io/kubernetes/vendor/k8s.io/apimachinery/pkg/util/wait.PollImmediateWithContext({0x7fe0bc8, 0xc0000820c8}, 0xc0031e9b60?, 0xc003475bb8?, 0x262a967?) vendor/k8s.io/apimachinery/pkg/util/wait/wait.go:528 k8s.io/kubernetes/vendor/k8s.io/apimachinery/pkg/util/wait.PollImmediate(0xc000dd2130?, 0x7fa7740?, 0xc000202b80?) vendor/k8s.io/apimachinery/pkg/util/wait/wait.go:514 k8s.io/kubernetes/test/e2e/framework/service.(*TestJig).waitForCondition(0xc003b567d0, 0x4?, {0x7600fe2, 0x14}, 0x7895b68) test/e2e/framework/service/jig.go:631 k8s.io/kubernetes/test/e2e/framework/service.(*TestJig).WaitForLoadBalancer(0xc003b567d0, 0x33?) test/e2e/framework/service/jig.go:582 > k8s.io/kubernetes/test/e2e/network.glob..func19.4() test/e2e/network/loadbalancer.go:381 k8s.io/kubernetes/vendor/github.com/onsi/ginkgo/v2/internal.extractBodyFunction.func3({0x2d591ce, 0xc0056e0000}) vendor/github.com/onsi/ginkgo/v2/internal/node.go:449 k8s.io/kubernetes/vendor/github.com/onsi/ginkgo/v2/internal.(*Suite).runNode.func2() vendor/github.com/onsi/ginkgo/v2/internal/suite.go:750 k8s.io/kubernetes/vendor/github.com/onsi/ginkgo/v2/internal.(*Suite).runNode vendor/github.com/onsi/ginkgo/v2/internal/suite.go:738 ------------------------------ Nov 26 05:38:07.079: INFO: UDP load balancer: 35.247.7.127 STEP: hitting the UDP service's NodePort 11/26/22 05:38:07.079 Nov 26 05:38:07.079: INFO: Poking udp://35.247.104.236:31435 Nov 26 05:38:07.119: INFO: Poke("udp://35.247.104.236:31435"): read udp 10.60.29.184:38091->35.247.104.236:31435: read: connection refused Nov 26 05:38:09.120: INFO: Poking udp://35.247.104.236:31435 Nov 26 05:38:09.159: INFO: Poke("udp://35.247.104.236:31435"): read udp 10.60.29.184:42199->35.247.104.236:31435: read: connection refused Nov 26 05:38:11.120: INFO: Poking udp://35.247.104.236:31435 Nov 26 05:38:11.159: INFO: Poke("udp://35.247.104.236:31435"): read udp 10.60.29.184:52524->35.247.104.236:31435: read: connection refused Nov 26 05:38:13.119: INFO: Poking udp://35.247.104.236:31435 Nov 26 05:38:13.158: INFO: Poke("udp://35.247.104.236:31435"): read udp 10.60.29.184:57302->35.247.104.236:31435: read: connection refused Nov 26 05:38:15.119: INFO: Poking udp://35.247.104.236:31435 Nov 26 05:38:15.158: INFO: Poke("udp://35.247.104.236:31435"): read udp 10.60.29.184:48075->35.247.104.236:31435: read: connection refused Nov 26 05:38:17.120: INFO: Poking udp://35.247.104.236:31435 Nov 26 05:38:17.159: INFO: Poke("udp://35.247.104.236:31435"): read udp 10.60.29.184:58222->35.247.104.236:31435: read: connection refused Nov 26 05:38:19.120: INFO: Poking udp://35.247.104.236:31435 Nov 26 05:38:19.159: INFO: Poke("udp://35.247.104.236:31435"): read udp 10.60.29.184:44922->35.247.104.236:31435: read: connection refused ------------------------------ Progress Report for Ginkgo Process #6 Automatically polling progress: [sig-network] LoadBalancers should be able to change the type and ports of a UDP service [Slow] (Spec Runtime: 14m40.645s) test/e2e/network/loadbalancer.go:287 In [It] (Node Runtime: 14m40.083s) test/e2e/network/loadbalancer.go:287 At [By Step] hitting the UDP service's NodePort (Step Runtime: 13.636s) test/e2e/network/loadbalancer.go:389 Spec Goroutine goroutine 1768 [select] k8s.io/kubernetes/vendor/k8s.io/apimachinery/pkg/util/wait.WaitForWithContext({0x7fe0bc8, 0xc0000820c8}, 0xc00165d188, 0x2fdb16a?) vendor/k8s.io/apimachinery/pkg/util/wait/wait.go:660 k8s.io/kubernetes/vendor/k8s.io/apimachinery/pkg/util/wait.poll({0x7fe0bc8, 0xc0000820c8}, 0x60?, 0x2fd9d05?, 0x10?) vendor/k8s.io/apimachinery/pkg/util/wait/wait.go:596 k8s.io/kubernetes/vendor/k8s.io/apimachinery/pkg/util/wait.PollImmediateWithContext({0x7fe0bc8, 0xc0000820c8}, 0xc000202b80?, 0xc0046c5cb0?, 0x262a967?) vendor/k8s.io/apimachinery/pkg/util/wait/wait.go:528 k8s.io/kubernetes/vendor/k8s.io/apimachinery/pkg/util/wait.PollImmediate(0x65cbc00?, 0xc0002019c8?, 0x754e980?) vendor/k8s.io/apimachinery/pkg/util/wait/wait.go:514 > k8s.io/kubernetes/test/e2e/network.testReachableUDP({0xc004c3de70, 0xe}, 0x7acb, 0x0?) test/e2e/network/service.go:603 > k8s.io/kubernetes/test/e2e/network.glob..func19.4() test/e2e/network/loadbalancer.go:390 k8s.io/kubernetes/vendor/github.com/onsi/ginkgo/v2/internal.extractBodyFunction.func3({0x2d591ce, 0xc0056e0000}) vendor/github.com/onsi/ginkgo/v2/internal/node.go:449 k8s.io/kubernetes/vendor/github.com/onsi/ginkgo/v2/internal.(*Suite).runNode.func2() vendor/github.com/onsi/ginkgo/v2/internal/suite.go:750 k8s.io/kubernetes/vendor/github.com/onsi/ginkgo/v2/internal.(*Suite).runNode vendor/github.com/onsi/ginkgo/v2/internal/suite.go:738 ------------------------------ Nov 26 05:38:21.119: INFO: Poking udp://35.247.104.236:31435 Nov 26 05:38:21.158: INFO: Poke("udp://35.247.104.236:31435"): read udp 10.60.29.184:46950->35.247.104.236:31435: read: connection refused Nov 26 05:38:23.120: INFO: Poking udp://35.247.104.236:31435 Nov 26 05:38:23.159: INFO: Poke("udp://35.247.104.236:31435"): read udp 10.60.29.184:55617->35.247.104.236:31435: read: connection refused Nov 26 05:38:25.120: INFO: Poking udp://35.247.104.236:31435 Nov 26 05:38:25.159: INFO: Poke("udp://35.247.104.236:31435"): read udp 10.60.29.184:42640->35.247.104.236:31435: read: connection refused Nov 26 05:38:27.120: INFO: Poking udp://35.247.104.236:31435 Nov 26 05:38:27.159: INFO: Poke("udp://35.247.104.236:31435"): read udp 10.60.29.184:39145->35.247.104.236:31435: read: connection refused Nov 26 05:38:29.120: INFO: Poking udp://35.247.104.236:31435 Nov 26 05:38:29.159: INFO: Poke("udp://35.247.104.236:31435"): read udp 10.60.29.184:39624->35.247.104.236:31435: read: connection refused Nov 26 05:38:31.120: INFO: Poking udp://35.247.104.236:31435 Nov 26 05:38:31.159: INFO: Poke("udp://35.247.104.236:31435"): read udp 10.60.29.184:40615->35.247.104.236:31435: read: connection refused Nov 26 05:38:33.120: INFO: Poking udp://35.247.104.236:31435 Nov 26 05:38:33.159: INFO: Poke("udp://35.247.104.236:31435"): read udp 10.60.29.184:52522->35.247.104.236:31435: read: connection refused Nov 26 05:38:35.119: INFO: Poking udp://35.247.104.236:31435 Nov 26 05:38:35.158: INFO: Poke("udp://35.247.104.236:31435"): read udp 10.60.29.184:59670->35.247.104.236:31435: read: connection refused Nov 26 05:38:37.119: INFO: Poking udp://35.247.104.236:31435 Nov 26 05:38:37.158: INFO: Poke("udp://35.247.104.236:31435"): read udp 10.60.29.184:41970->35.247.104.236:31435: read: connection refused Nov 26 05:38:39.120: INFO: Poking udp://35.247.104.236:31435 Nov 26 05:38:39.159: INFO: Poke("udp://35.247.104.236:31435"): read udp 10.60.29.184:56151->35.247.104.236:31435: read: connection refused ------------------------------ Progress Report for Ginkgo Process #6 Automatically polling progress: [sig-network] LoadBalancers should be able to change the type and ports of a UDP service [Slow] (Spec Runtime: 15m0.648s) test/e2e/network/loadbalancer.go:287 In [It] (Node Runtime: 15m0.085s) test/e2e/network/loadbalancer.go:287 At [By Step] hitting the UDP service's NodePort (Step Runtime: 33.638s) test/e2e/network/loadbalancer.go:389 Spec Goroutine goroutine 1768 [select] k8s.io/kubernetes/vendor/k8s.io/apimachinery/pkg/util/wait.WaitForWithContext({0x7fe0bc8, 0xc0000820c8}, 0xc00165d188, 0x2fdb16a?) vendor/k8s.io/apimachinery/pkg/util/wait/wait.go:660 k8s.io/kubernetes/vendor/k8s.io/apimachinery/pkg/util/wait.poll({0x7fe0bc8, 0xc0000820c8}, 0x60?, 0x2fd9d05?, 0x10?) vendor/k8s.io/apimachinery/pkg/util/wait/wait.go:596 k8s.io/kubernetes/vendor/k8s.io/apimachinery/pkg/util/wait.PollImmediateWithContext({0x7fe0bc8, 0xc0000820c8}, 0xc000202b80?, 0xc0046c5cb0?, 0x262a967?) vendor/k8s.io/apimachinery/pkg/util/wait/wait.go:528 k8s.io/kubernetes/vendor/k8s.io/apimachinery/pkg/util/wait.PollImmediate(0x65cbc00?, 0xc0002019c8?, 0x754e980?) vendor/k8s.io/apimachinery/pkg/util/wait/wait.go:514 > k8s.io/kubernetes/test/e2e/network.testReachableUDP({0xc004c3de70, 0xe}, 0x7acb, 0x0?) test/e2e/network/service.go:603 > k8s.io/kubernetes/test/e2e/network.glob..func19.4() test/e2e/network/loadbalancer.go:390 k8s.io/kubernetes/vendor/github.com/onsi/ginkgo/v2/internal.extractBodyFunction.func3({0x2d591ce, 0xc0056e0000}) vendor/github.com/onsi/ginkgo/v2/internal/node.go:449 k8s.io/kubernetes/vendor/github.com/onsi/ginkgo/v2/internal.(*Suite).runNode.func2() vendor/github.com/onsi/ginkgo/v2/internal/suite.go:750 k8s.io/kubernetes/vendor/github.com/onsi/ginkgo/v2/internal.(*Suite).runNode vendor/github.com/onsi/ginkgo/v2/internal/suite.go:738 ------------------------------ Nov 26 05:38:41.120: INFO: Poking udp://35.247.104.236:31435 Nov 26 05:38:41.159: INFO: Poke("udp://35.247.104.236:31435"): read udp 10.60.29.184:51400->35.247.104.236:31435: read: connection refused Nov 26 05:38:43.120: INFO: Poking udp://35.247.104.236:31435 Nov 26 05:38:43.159: INFO: Poke("udp://35.247.104.236:31435"): read udp 10.60.29.184:40954->35.247.104.236:31435: read: connection refused Nov 26 05:38:45.120: INFO: Poking udp://35.247.104.236:31435 Nov 26 05:38:45.159: INFO: Poke("udp://35.247.104.236:31435"): read udp 10.60.29.184:55101->35.247.104.236:31435: read: connection refused Nov 26 05:38:47.120: INFO: Poking udp://35.247.104.236:31435 Nov 26 05:38:47.159: INFO: Poke("udp://35.247.104.236:31435"): read udp 10.60.29.184:44398->35.247.104.236:31435: read: connection refused Nov 26 05:38:49.120: INFO: Poking udp://35.247.104.236:31435 Nov 26 05:38:49.159: INFO: Poke("udp://35.247.104.236:31435"): read udp 10.60.29.184:42541->35.247.104.236:31435: read: connection refused Nov 26 05:38:51.119: INFO: Poking udp://35.247.104.236:31435 Nov 26 05:38:51.158: INFO: Poke("udp://35.247.104.236:31435"): read udp 10.60.29.184:48452->35.247.104.236:31435: read: connection refused Nov 26 05:38:53.119: INFO: Poking udp://35.247.104.236:31435 Nov 26 05:38:53.158: INFO: Poke("udp://35.247.104.236:31435"): read udp 10.60.29.184:52527->35.247.104.236:31435: read: connection refused Nov 26 05:38:55.120: INFO: Poking udp://35.247.104.236:31435 Nov 26 05:38:55.159: INFO: Poke("udp://35.247.104.236:31435"): read udp 10.60.29.184:50089->35.247.104.236:31435: read: connection refused Nov 26 05:38:57.120: INFO: Poking udp://35.247.104.236:31435 Nov 26 05:38:57.159: INFO: Poke("udp://35.247.104.236:31435"): read udp 10.60.29.184:36226->35.247.104.236:31435: read: connection refused Nov 26 05:38:59.119: INFO: Poking udp://35.247.104.236:31435 Nov 26 05:38:59.159: INFO: Poke("udp://35.247.104.236:31435"): read udp 10.60.29.184:49894->35.247.104.236:31435: read: connection refused ------------------------------ Progress Report for Ginkgo Process #6 Automatically polling progress: [sig-network] LoadBalancers should be able to change the type and ports of a UDP service [Slow] (Spec Runtime: 15m20.649s) test/e2e/network/loadbalancer.go:287 In [It] (Node Runtime: 15m20.087s) test/e2e/network/loadbalancer.go:287 At [By Step] hitting the UDP service's NodePort (Step Runtime: 53.64s) test/e2e/network/loadbalancer.go:389 Spec Goroutine goroutine 1768 [select] k8s.io/kubernetes/vendor/k8s.io/apimachinery/pkg/util/wait.WaitForWithContext({0x7fe0bc8, 0xc0000820c8}, 0xc00165d188, 0x2fdb16a?) vendor/k8s.io/apimachinery/pkg/util/wait/wait.go:660 k8s.io/kubernetes/vendor/k8s.io/apimachinery/pkg/util/wait.poll({0x7fe0bc8, 0xc0000820c8}, 0x60?, 0x2fd9d05?, 0x10?) vendor/k8s.io/apimachinery/pkg/util/wait/wait.go:596 k8s.io/kubernetes/vendor/k8s.io/apimachinery/pkg/util/wait.PollImmediateWithContext({0x7fe0bc8, 0xc0000820c8}, 0xc000202b80?, 0xc0046c5cb0?, 0x262a967?) vendor/k8s.io/apimachinery/pkg/util/wait/wait.go:528 k8s.io/kubernetes/vendor/k8s.io/apimachinery/pkg/util/wait.PollImmediate(0x65cbc00?, 0xc0002019c8?, 0x754e980?) vendor/k8s.io/apimachinery/pkg/util/wait/wait.go:514 > k8s.io/kubernetes/test/e2e/network.testReachableUDP({0xc004c3de70, 0xe}, 0x7acb, 0x0?) test/e2e/network/service.go:603 > k8s.io/kubernetes/test/e2e/network.glob..func19.4() test/e2e/network/loadbalancer.go:390 k8s.io/kubernetes/vendor/github.com/onsi/ginkgo/v2/internal.extractBodyFunction.func3({0x2d591ce, 0xc0056e0000}) vendor/github.com/onsi/ginkgo/v2/internal/node.go:449 k8s.io/kubernetes/vendor/github.com/onsi/ginkgo/v2/internal.(*Suite).runNode.func2() vendor/github.com/onsi/ginkgo/v2/internal/suite.go:750 k8s.io/kubernetes/vendor/github.com/onsi/ginkgo/v2/internal.(*Suite).runNode vendor/github.com/onsi/ginkgo/v2/internal/suite.go:738 ------------------------------ Nov 26 05:39:01.120: INFO: Poking udp://35.247.104.236:31435 Nov 26 05:39:01.159: INFO: Poke("udp://35.247.104.236:31435"): read udp 10.60.29.184:41941->35.247.104.236:31435: read: connection refused Nov 26 05:39:03.120: INFO: Poking udp://35.247.104.236:31435 Nov 26 05:39:03.159: INFO: Poke("udp://35.247.104.236:31435"): read udp 10.60.29.184:57221->35.247.104.236:31435: read: connection refused Nov 26 05:39:05.119: INFO: Poking udp://35.247.104.236:31435 Nov 26 05:39:05.158: INFO: Poke("udp://35.247.104.236:31435"): read udp 10.60.29.184:38472->35.247.104.236:31435: read: connection refused Nov 26 05:39:07.120: INFO: Poking udp://35.247.104.236:31435 Nov 26 05:39:07.159: INFO: Poke("udp://35.247.104.236:31435"): read udp 10.60.29.184:37551->35.247.104.236:31435: read: connection refused Nov 26 05:39:09.119: INFO: Poking udp://35.247.104.236:31435 Nov 26 05:39:09.158: INFO: Poke("udp://35.247.104.236:31435"): read udp 10.60.29.184:55519->35.247.104.236:31435: read: connection refused Nov 26 05:39:11.120: INFO: Poking udp://35.247.104.236:31435 Nov 26 05:39:11.159: INFO: Poke("udp://35.247.104.236:31435"): read udp 10.60.29.184:40985->35.247.104.236:31435: read: connection refused Nov 26 05:39:13.119: INFO: Poking udp://35.247.104.236:31435 Nov 26 05:39:13.158: INFO: Poke("udp://35.247.104.236:31435"): read udp 10.60.29.184:33810->35.247.104.236:31435: read: connection refused Nov 26 05:39:15.119: INFO: Poking udp://35.247.104.236:31435 Nov 26 05:39:15.158: INFO: Poke("udp://35.247.104.236:31435"): read udp 10.60.29.184:34825->35.247.104.236:31435: read: connection refused Nov 26 05:39:17.119: INFO: Poking udp://35.247.104.236:31435 Nov 26 05:39:17.158: INFO: Poke("udp://35.247.104.236:31435"): read udp 10.60.29.184:34444->35.247.104.236:31435: read: connection refused Nov 26 05:39:19.120: INFO: Poking udp://35.247.104.236:31435 Nov 26 05:39:19.159: INFO: Poke("udp://35.247.104.236:31435"): read udp 10.60.29.184:44247->35.247.104.236:31435: read: connection refused ------------------------------ Progress Report for Ginkgo Process #6 Automatically polling progress: [sig-network] LoadBalancers should be able to change the type and ports of a UDP service [Slow] (Spec Runtime: 15m40.651s) test/e2e/network/loadbalancer.go:287 In [It] (Node Runtime: 15m40.088s) test/e2e/network/loadbalancer.go:287 At [By Step] hitting the UDP service's NodePort (Step Runtime: 1m13.642s) test/e2e/network/loadbalancer.go:389 Spec Goroutine goroutine 1768 [select] k8s.io/kubernetes/vendor/k8s.io/apimachinery/pkg/util/wait.WaitForWithContext({0x7fe0bc8, 0xc0000820c8}, 0xc00165d188, 0x2fdb16a?) vendor/k8s.io/apimachinery/pkg/util/wait/wait.go:660 k8s.io/kubernetes/vendor/k8s.io/apimachinery/pkg/util/wait.poll({0x7fe0bc8, 0xc0000820c8}, 0x60?, 0x2fd9d05?, 0x10?) vendor/k8s.io/apimachinery/pkg/util/wait/wait.go:596 k8s.io/kubernetes/vendor/k8s.io/apimachinery/pkg/util/wait.PollImmediateWithContext({0x7fe0bc8, 0xc0000820c8}, 0xc000202b80?, 0xc0046c5cb0?, 0x262a967?) vendor/k8s.io/apimachinery/pkg/util/wait/wait.go:528 k8s.io/kubernetes/vendor/k8s.io/apimachinery/pkg/util/wait.PollImmediate(0x65cbc00?, 0xc0002019c8?, 0x754e980?) vendor/k8s.io/apimachinery/pkg/util/wait/wait.go:514 > k8s.io/kubernetes/test/e2e/network.testReachableUDP({0xc004c3de70, 0xe}, 0x7acb, 0x0?) test/e2e/network/service.go:603 > k8s.io/kubernetes/test/e2e/network.glob..func19.4() test/e2e/network/loadbalancer.go:390 k8s.io/kubernetes/vendor/github.com/onsi/ginkgo/v2/internal.extractBodyFunction.func3({0x2d591ce, 0xc0056e0000}) vendor/github.com/onsi/ginkgo/v2/internal/node.go:449 k8s.io/kubernetes/vendor/github.com/onsi/ginkgo/v2/internal.(*Suite).runNode.func2() vendor/github.com/onsi/ginkgo/v2/internal/suite.go:750 k8s.io/kubernetes/vendor/github.com/onsi/ginkgo/v2/internal.(*Suite).runNode vendor/github.com/onsi/ginkgo/v2/internal/suite.go:738 ------------------------------ Nov 26 05:39:21.120: INFO: Poking udp://35.247.104.236:31435 Nov 26 05:39:21.159: INFO: Poke("udp://35.247.104.236:31435"): read udp 10.60.29.184:55407->35.247.104.236:31435: read: connection refused Nov 26 05:39:23.120: INFO: Poking udp://35.247.104.236:31435 Nov 26 05:39:23.159: INFO: Poke("udp://35.247.104.236:31435"): read udp 10.60.29.184:35519->35.247.104.236:31435: read: connection refused Nov 26 05:39:25.119: INFO: Poking udp://35.247.104.236:31435 Nov 26 05:39:25.158: INFO: Poke("udp://35.247.104.236:31435"): read udp 10.60.29.184:48118->35.247.104.236:31435: read: connection refused Nov 26 05:39:27.120: INFO: Poking udp://35.247.104.236:31435 Nov 26 05:39:27.159: INFO: Poke("udp://35.247.104.236:31435"): read udp 10.60.29.184:56546->35.247.104.236:31435: read: connection refused Nov 26 05:39:29.120: INFO: Poking udp://35.247.104.236:31435 Nov 26 05:39:29.159: INFO: Poke("udp://35.247.104.236:31435"): read udp 10.60.29.184:44892->35.247.104.236:31435: read: connection refused Nov 26 05:39:31.119: INFO: Poking udp://35.247.104.236:31435 Nov 26 05:39:31.158: INFO: Poke("udp://35.247.104.236:31435"): read udp 10.60.29.184:39190->35.247.104.236:31435: read: connection refused Nov 26 05:39:33.119: INFO: Poking udp://35.247.104.236:31435 Nov 26 05:39:33.158: INFO: Poke("udp://35.247.104.236:31435"): read udp 10.60.29.184:33158->35.247.104.236:31435: read: connection refused Nov 26 05:39:35.120: INFO: Poking udp://35.247.104.236:31435 Nov 26 05:39:35.159: INFO: Poke("udp://35.247.104.236:31435"): read udp 10.60.29.184:45163->35.247.104.236:31435: read: connection refused Nov 26 05:39:37.119: INFO: Poking udp://35.247.104.236:31435 Nov 26 05:39:37.159: INFO: Poke("udp://35.247.104.236:31435"): read udp 10.60.29.184:44934->35.247.104.236:31435: read: connection refused Nov 26 05:39:39.119: INFO: Poking udp://35.247.104.236:31435 Nov 26 05:39:39.158: INFO: Poke("udp://35.247.104.236:31435"): read udp 10.60.29.184:34266->35.247.104.236:31435: read: connection refused ------------------------------ Progress Report for Ginkgo Process #6 Automatically polling progress: [sig-network] LoadBalancers should be able to change the type and ports of a UDP service [Slow] (Spec Runtime: 16m0.653s) test/e2e/network/loadbalancer.go:287 In [It] (Node Runtime: 16m0.09s) test/e2e/network/loadbalancer.go:287 At [By Step] hitting the UDP service's NodePort (Step Runtime: 1m33.643s) test/e2e/network/loadbalancer.go:389 Spec Goroutine goroutine 1768 [select] k8s.io/kubernetes/vendor/k8s.io/apimachinery/pkg/util/wait.WaitForWithContext({0x7fe0bc8, 0xc0000820c8}, 0xc00165d188, 0x2fdb16a?) vendor/k8s.io/apimachinery/pkg/util/wait/wait.go:660 k8s.io/kubernetes/vendor/k8s.io/apimachinery/pkg/util/wait.poll({0x7fe0bc8, 0xc0000820c8}, 0x60?, 0x2fd9d05?, 0x10?) vendor/k8s.io/apimachinery/pkg/util/wait/wait.go:596 k8s.io/kubernetes/vendor/k8s.io/apimachinery/pkg/util/wait.PollImmediateWithContext({0x7fe0bc8, 0xc0000820c8}, 0xc000202b80?, 0xc0046c5cb0?, 0x262a967?) vendor/k8s.io/apimachinery/pkg/util/wait/wait.go:528 k8s.io/kubernetes/vendor/k8s.io/apimachinery/pkg/util/wait.PollImmediate(0x65cbc00?, 0xc0002019c8?, 0x754e980?) vendor/k8s.io/apimachinery/pkg/util/wait/wait.go:514 > k8s.io/kubernetes/test/e2e/network.testReachableUDP({0xc004c3de70, 0xe}, 0x7acb, 0x0?) test/e2e/network/service.go:603 > k8s.io/kubernetes/test/e2e/network.glob..func19.4() test/e2e/network/loadbalancer.go:390 k8s.io/kubernetes/vendor/github.com/onsi/ginkgo/v2/internal.extractBodyFunction.func3({0x2d591ce, 0xc0056e0000}) vendor/github.com/onsi/ginkgo/v2/internal/node.go:449 k8s.io/kubernetes/vendor/github.com/onsi/ginkgo/v2/internal.(*Suite).runNode.func2() vendor/github.com/onsi/ginkgo/v2/internal/suite.go:750 k8s.io/kubernetes/vendor/github.com/onsi/ginkgo/v2/internal.(*Suite).runNode vendor/github.com/onsi/ginkgo/v2/internal/suite.go:738 ------------------------------ Nov 26 05:39:41.120: INFO: Poking udp://35.247.104.236:31435 Nov 26 05:39:41.159: INFO: Poke("udp://35.247.104.236:31435"): read udp 10.60.29.184:56838->35.247.104.236:31435: read: connection refused Nov 26 05:39:43.120: INFO: Poking udp://35.247.104.236:31435 Nov 26 05:39:43.159: INFO: Poke("udp://35.247.104.236:31435"): read udp 10.60.29.184:42951->35.247.104.236:31435: read: connection refused Nov 26 05:39:45.120: INFO: Poking udp://35.247.104.236:31435 Nov 26 05:39:45.159: INFO: Poke("udp://35.247.104.236:31435"): read udp 10.60.29.184:53991->35.247.104.236:31435: read: connection refused Nov 26 05:39:47.120: INFO: Poking udp://35.247.104.236:31435 Nov 26 05:39:47.159: INFO: Poke("udp://35.247.104.236:31435"): read udp 10.60.29.184:44866->35.247.104.236:31435: read: connection refused Nov 26 05:39:49.120: INFO: Poking udp://35.247.104.236:31435 Nov 26 05:39:49.159: INFO: Poke("udp://35.247.104.236:31435"): read udp 10.60.29.184:59895->35.247.104.236:31435: read: connection refused Nov 26 05:39:51.120: INFO: Poking udp://35.247.104.236:31435 Nov 26 05:39:51.159: INFO: Poke("udp://35.247.104.236:31435"): read udp 10.60.29.184:54604->35.247.104.236:31435: read: connection refused Nov 26 05:39:53.120: INFO: Poking udp://35.247.104.236:31435 Nov 26 05:39:53.159: INFO: Poke("udp://35.247.104.236:31435"): read udp 10.60.29.184:39563->35.247.104.236:31435: read: connection refused Nov 26 05:39:55.119: INFO: Poking udp://35.247.104.236:31435 Nov 26 05:39:55.159: INFO: Poke("udp://35.247.104.236:31435"): read udp 10.60.29.184:42342->35.247.104.236:31435: read: connection refused Nov 26 05:39:57.120: INFO: Poking udp://35.247.104.236:31435 Nov 26 05:39:57.159: INFO: Poke("udp://35.247.104.236:31435"): read udp 10.60.29.184:34667->35.247.104.236:31435: read: connection refused Nov 26 05:39:59.120: INFO: Poking udp://35.247.104.236:31435 Nov 26 05:39:59.159: INFO: Poke("udp://35.247.104.236:31435"): read udp 10.60.29.184:55633->35.247.104.236:31435: read: connection refused ------------------------------ Progress Report for Ginkgo Process #6 Automatically polling progress: [sig-network] LoadBalancers should be able to change the type and ports of a UDP service [Slow] (Spec Runtime: 16m20.654s) test/e2e/network/loadbalancer.go:287 In [It] (Node Runtime: 16m20.092s) test/e2e/network/loadbalancer.go:287 At [By Step] hitting the UDP service's NodePort (Step Runtime: 1m53.645s) test/e2e/network/loadbalancer.go:389 Spec Goroutine goroutine 1768 [select] k8s.io/kubernetes/vendor/k8s.io/apimachinery/pkg/util/wait.WaitForWithContext({0x7fe0bc8, 0xc0000820c8}, 0xc00165d188, 0x2fdb16a?) vendor/k8s.io/apimachinery/pkg/util/wait/wait.go:660 k8s.io/kubernetes/vendor/k8s.io/apimachinery/pkg/util/wait.poll({0x7fe0bc8, 0xc0000820c8}, 0x60?, 0x2fd9d05?, 0x10?) vendor/k8s.io/apimachinery/pkg/util/wait/wait.go:596 k8s.io/kubernetes/vendor/k8s.io/apimachinery/pkg/util/wait.PollImmediateWithContext({0x7fe0bc8, 0xc0000820c8}, 0xc000202b80?, 0xc0046c5cb0?, 0x262a967?) vendor/k8s.io/apimachinery/pkg/util/wait/wait.go:528 k8s.io/kubernetes/vendor/k8s.io/apimachinery/pkg/util/wait.PollImmediate(0x65cbc00?, 0xc0002019c8?, 0x754e980?) vendor/k8s.io/apimachinery/pkg/util/wait/wait.go:514 > k8s.io/kubernetes/test/e2e/network.testReachableUDP({0xc004c3de70, 0xe}, 0x7acb, 0x0?) test/e2e/network/service.go:603 > k8s.io/kubernetes/test/e2e/network.glob..func19.4() test/e2e/network/loadbalancer.go:390 k8s.io/kubernetes/vendor/github.com/onsi/ginkgo/v2/internal.extractBodyFunction.func3({0x2d591ce, 0xc0056e0000}) vendor/github.com/onsi/ginkgo/v2/internal/node.go:449 k8s.io/kubernetes/vendor/github.com/onsi/ginkgo/v2/internal.(*Suite).runNode.func2() vendor/github.com/onsi/ginkgo/v2/internal/suite.go:750 k8s.io/kubernetes/vendor/github.com/onsi/ginkgo/v2/internal.(*Suite).runNode vendor/github.com/onsi/ginkgo/v2/internal/suite.go:738 ------------------------------ Nov 26 05:40:01.120: INFO: Poking udp://35.247.104.236:31435 Nov 26 05:40:01.159: INFO: Poke("udp://35.247.104.236:31435"): read udp 10.60.29.184:55987->35.247.104.236:31435: read: connection refused Nov 26 05:40:03.119: INFO: Poking udp://35.247.104.236:31435 Nov 26 05:40:03.159: INFO: Poke("udp://35.247.104.236:31435"): read udp 10.60.29.184:33138->35.247.104.236:31435: read: connection refused Nov 26 05:40:05.119: INFO: Poking udp://35.247.104.236:31435 Nov 26 05:40:05.159: INFO: Poke("udp://35.247.104.236:31435"): read udp 10.60.29.184:48547->35.247.104.236:31435: read: connection refused Nov 26 05:40:07.120: INFO: Poking udp://35.247.104.236:31435 Nov 26 05:40:07.159: INFO: Poke("udp://35.247.104.236:31435"): read udp 10.60.29.184:53547->35.247.104.236:31435: read: connection refused Nov 26 05:40:09.120: INFO: Poking udp://35.247.104.236:31435 Nov 26 05:40:09.159: INFO: Poke("udp://35.247.104.236:31435"): read udp 10.60.29.184:41179->35.247.104.236:31435: read: connection refused Nov 26 05:40:11.120: INFO: Poking udp://35.247.104.236:31435 Nov 26 05:40:11.159: INFO: Poke("udp://35.247.104.236:31435"): read udp 10.60.29.184:37142->35.247.104.236:31435: read: connection refused Nov 26 05:40:13.120: INFO: Poking udp://35.247.104.236:31435 Nov 26 05:40:13.159: INFO: Poke("udp://35.247.104.236:31435"): read udp 10.60.29.184:38989->35.247.104.236:31435: read: connection refused Nov 26 05:40:15.119: INFO: Poking udp://35.247.104.236:31435 Nov 26 05:40:15.158: INFO: Poke("udp://35.247.104.236:31435"): read udp 10.60.29.184:54818->35.247.104.236:31435: read: connection refused Nov 26 05:40:17.120: INFO: Poking udp://35.247.104.236:31435 Nov 26 05:40:17.159: INFO: Poke("udp://35.247.104.236:31435"): read udp 10.60.29.184:33672->35.247.104.236:31435: read: connection refused Nov 26 05:40:19.120: INFO: Poking udp://35.247.104.236:31435 Nov 26 05:40:19.159: INFO: Poke("udp://35.247.104.236:31435"): read udp 10.60.29.184:53369->35.247.104.236:31435: read: connection refused ------------------------------ Progress Report for Ginkgo Process #6 Automatically polling progress: [sig-network] LoadBalancers should be able to change the type and ports of a UDP service [Slow] (Spec Runtime: 16m40.656s) test/e2e/network/loadbalancer.go:287 In [It] (Node Runtime: 16m40.093s) test/e2e/network/loadbalancer.go:287 At [By Step] hitting the UDP service's NodePort (Step Runtime: 2m13.646s) test/e2e/network/loadbalancer.go:389 Spec Goroutine goroutine 1768 [select] k8s.io/kubernetes/vendor/k8s.io/apimachinery/pkg/util/wait.WaitForWithContext({0x7fe0bc8, 0xc0000820c8}, 0xc00165d188, 0x2fdb16a?) vendor/k8s.io/apimachinery/pkg/util/wait/wait.go:660 k8s.io/kubernetes/vendor/k8s.io/apimachinery/pkg/util/wait.poll({0x7fe0bc8, 0xc0000820c8}, 0x60?, 0x2fd9d05?, 0x10?) vendor/k8s.io/apimachinery/pkg/util/wait/wait.go:596 k8s.io/kubernetes/vendor/k8s.io/apimachinery/pkg/util/wait.PollImmediateWithContext({0x7fe0bc8, 0xc0000820c8}, 0xc000202b80?, 0xc0046c5cb0?, 0x262a967?) vendor/k8s.io/apimachinery/pkg/util/wait/wait.go:528 k8s.io/kubernetes/vendor/k8s.io/apimachinery/pkg/util/wait.PollImmediate(0x65cbc00?, 0xc0002019c8?, 0x754e980?) vendor/k8s.io/apimachinery/pkg/util/wait/wait.go:514 > k8s.io/kubernetes/test/e2e/network.testReachableUDP({0xc004c3de70, 0xe}, 0x7acb, 0x0?) test/e2e/network/service.go:603 > k8s.io/kubernetes/test/e2e/network.glob..func19.4() test/e2e/network/loadbalancer.go:390 k8s.io/kubernetes/vendor/github.com/onsi/ginkgo/v2/internal.extractBodyFunction.func3({0x2d591ce, 0xc0056e0000}) vendor/github.com/onsi/ginkgo/v2/internal/node.go:449 k8s.io/kubernetes/vendor/github.com/onsi/ginkgo/v2/internal.(*Suite).runNode.func2() vendor/github.com/onsi/ginkgo/v2/internal/suite.go:750 k8s.io/kubernetes/vendor/github.com/onsi/ginkgo/v2/internal.(*Suite).runNode vendor/github.com/onsi/ginkgo/v2/internal/suite.go:738 ------------------------------ Nov 26 05:40:21.120: INFO: Poking udp://35.247.104.236:31435 Nov 26 05:40:21.159: INFO: Poke("udp://35.247.104.236:31435"): read udp 10.60.29.184:33437->35.247.104.236:31435: read: connection refused Nov 26 05:40:23.120: INFO: Poking udp://35.247.104.236:31435 Nov 26 05:40:23.159: INFO: Poke("udp://35.247.104.236:31435"): read udp 10.60.29.184:48929->35.247.104.236:31435: read: connection refused Nov 26 05:40:25.119: INFO: Poking udp://35.247.104.236:31435 Nov 26 05:40:25.158: INFO: Poke("udp://35.247.104.236:31435"): read udp 10.60.29.184:49035->35.247.104.236:31435: read: connection refused Nov 26 05:40:27.120: INFO: Poking udp://35.247.104.236:31435 Nov 26 05:40:27.159: INFO: Poke("udp://35.247.104.236:31435"): read udp 10.60.29.184:48586->35.247.104.236:31435: read: connection refused Nov 26 05:40:29.120: INFO: Poking udp://35.247.104.236:31435 Nov 26 05:40:29.159: INFO: Poke("udp://35.247.104.236:31435"): read udp 10.60.29.184:38601->35.247.104.236:31435: read: connection refused Nov 26 05:40:31.120: INFO: Poking udp://35.247.104.236:31435 Nov 26 05:40:31.159: INFO: Poke("udp://35.247.104.236:31435"): read udp 10.60.29.184:40957->35.247.104.236:31435: read: connection refused Nov 26 05:40:33.120: INFO: Poking udp://35.247.104.236:31435 Nov 26 05:40:33.159: INFO: Poke("udp://35.247.104.236:31435"): read udp 10.60.29.184:38298->35.247.104.236:31435: read: connection refused Nov 26 05:40:35.119: INFO: Poking udp://35.247.104.236:31435 Nov 26 05:40:35.159: INFO: Poke("udp://35.247.104.236:31435"): read udp 10.60.29.184:56937->35.247.104.236:31435: read: connection refused Nov 26 05:40:37.120: INFO: Poking udp://35.247.104.236:31435 Nov 26 05:40:37.159: INFO: Poke("udp://35.247.104.236:31435"): read udp 10.60.29.184:57050->35.247.104.236:31435: read: connection refused Nov 26 05:40:39.120: INFO: Poking udp://35.247.104.236:31435 Nov 26 05:40:39.159: INFO: Poke("udp://35.247.104.236:31435"): read udp 10.60.29.184:52815->35.247.104.236:31435: read: connection refused ------------------------------ Progress Report for Ginkgo Process #6 Automatically polling progress: [sig-network] LoadBalancers should be able to change the type and ports of a UDP service [Slow] (Spec Runtime: 17m0.661s) test/e2e/network/loadbalancer.go:287 In [It] (Node Runtime: 17m0.099s) test/e2e/network/loadbalancer.go:287 At [By Step] hitting the UDP service's NodePort (Step Runtime: 2m33.652s) test/e2e/network/loadbalancer.go:389 Spec Goroutine goroutine 1768 [select] k8s.io/kubernetes/vendor/k8s.io/apimachinery/pkg/util/wait.WaitForWithContext({0x7fe0bc8, 0xc0000820c8}, 0xc00165d188, 0x2fdb16a?) vendor/k8s.io/apimachinery/pkg/util/wait/wait.go:660 k8s.io/kubernetes/vendor/k8s.io/apimachinery/pkg/util/wait.poll({0x7fe0bc8, 0xc0000820c8}, 0x60?, 0x2fd9d05?, 0x10?) vendor/k8s.io/apimachinery/pkg/util/wait/wait.go:596 k8s.io/kubernetes/vendor/k8s.io/apimachinery/pkg/util/wait.PollImmediateWithContext({0x7fe0bc8, 0xc0000820c8}, 0xc000202b80?, 0xc0046c5cb0?, 0x262a967?) vendor/k8s.io/apimachinery/pkg/util/wait/wait.go:528 k8s.io/kubernetes/vendor/k8s.io/apimachinery/pkg/util/wait.PollImmediate(0x65cbc00?, 0xc0002019c8?, 0x754e980?) vendor/k8s.io/apimachinery/pkg/util/wait/wait.go:514 > k8s.io/kubernetes/test/e2e/network.testReachableUDP({0xc004c3de70, 0xe}, 0x7acb, 0x0?) test/e2e/network/service.go:603 > k8s.io/kubernetes/test/e2e/network.glob..func19.4() test/e2e/network/loadbalancer.go:390 k8s.io/kubernetes/vendor/github.com/onsi/ginkgo/v2/internal.extractBodyFunction.func3({0x2d591ce, 0xc0056e0000}) vendor/github.com/onsi/ginkgo/v2/internal/node.go:449 k8s.io/kubernetes/vendor/github.com/onsi/ginkgo/v2/internal.(*Suite).runNode.func2() vendor/github.com/onsi/ginkgo/v2/internal/suite.go:750 k8s.io/kubernetes/vendor/github.com/onsi/ginkgo/v2/internal.(*Suite).runNode vendor/github.com/onsi/ginkgo/v2/internal/suite.go:738 ------------------------------ Nov 26 05:40:41.120: INFO: Poking udp://35.247.104.236:31435 Nov 26 05:40:41.159: INFO: Poke("udp://35.247.104.236:31435"): read udp 10.60.29.184:60858->35.247.104.236:31435: read: connection refused Nov 26 05:40:43.120: INFO: Poking udp://35.247.104.236:31435 Nov 26 05:40:43.159: INFO: Poke("udp://35.247.104.236:31435"): read udp 10.60.29.184:44915->35.247.104.236:31435: read: connection refused Nov 26 05:40:45.119: INFO: Poking udp://35.247.104.236:31435 Nov 26 05:40:45.158: INFO: Poke("udp://35.247.104.236:31435"): read udp 10.60.29.184:49248->35.247.104.236:31435: read: connection refused Nov 26 05:40:47.120: INFO: Poking udp://35.247.104.236:31435 Nov 26 05:40:47.173: INFO: Poke("udp://35.247.104.236:31435"): read udp 10.60.29.184:52159->35.247.104.236:31435: read: connection refused Nov 26 05:40:49.120: INFO: Poking udp://35.247.104.236:31435 Nov 26 05:40:49.159: INFO: Poke("udp://35.247.104.236:31435"): read udp 10.60.29.184:33442->35.247.104.236:31435: read: connection refused Nov 26 05:40:51.120: INFO: Poking udp://35.247.104.236:31435 Nov 26 05:40:51.159: INFO: Poke("udp://35.247.104.236:31435"): read udp 10.60.29.184:54359->35.247.104.236:31435: read: connection refused Nov 26 05:40:53.120: INFO: Poking udp://35.247.104.236:31435 Nov 26 05:40:53.159: INFO: Poke("udp://35.247.104.236:31435"): read udp 10.60.29.184:39328->35.247.104.236:31435: read: connection refused Nov 26 05:40:55.120: INFO: Poking udp://35.247.104.236:31435 Nov 26 05:40:55.159: INFO: Poke("udp://35.247.104.236:31435"): read udp 10.60.29.184:47068->35.247.104.236:31435: read: connection refused Nov 26 05:40:57.120: INFO: Poking udp://35.247.104.236:31435 Nov 26 05:40:57.160: INFO: Poke("udp://35.247.104.236:31435"): read udp 10.60.29.184:37650->35.247.104.236:31435: read: connection refused Nov 26 05:40:59.120: INFO: Poking udp://35.247.104.236:31435 Nov 26 05:40:59.159: INFO: Poke("udp://35.247.104.236:31435"): read udp 10.60.29.184:50094->35.247.104.236:31435: read: connection refused ------------------------------ Progress Report for Ginkgo Process #6 Automatically polling progress: [sig-network] LoadBalancers should be able to change the type and ports of a UDP service [Slow] (Spec Runtime: 17m20.663s) test/e2e/network/loadbalancer.go:287 In [It] (Node Runtime: 17m20.101s) test/e2e/network/loadbalancer.go:287 At [By Step] hitting the UDP service's NodePort (Step Runtime: 2m53.654s) test/e2e/network/loadbalancer.go:389 Spec Goroutine goroutine 1768 [select] k8s.io/kubernetes/vendor/k8s.io/apimachinery/pkg/util/wait.WaitForWithContext({0x7fe0bc8, 0xc0000820c8}, 0xc00165d188, 0x2fdb16a?) vendor/k8s.io/apimachinery/pkg/util/wait/wait.go:660 k8s.io/kubernetes/vendor/k8s.io/apimachinery/pkg/util/wait.poll({0x7fe0bc8, 0xc0000820c8}, 0x60?, 0x2fd9d05?, 0x10?) vendor/k8s.io/apimachinery/pkg/util/wait/wait.go:596 k8s.io/kubernetes/vendor/k8s.io/apimachinery/pkg/util/wait.PollImmediateWithContext({0x7fe0bc8, 0xc0000820c8}, 0xc000202b80?, 0xc0046c5cb0?, 0x262a967?) vendor/k8s.io/apimachinery/pkg/util/wait/wait.go:528 k8s.io/kubernetes/vendor/k8s.io/apimachinery/pkg/util/wait.PollImmediate(0x65cbc00?, 0xc0002019c8?, 0x754e980?) vendor/k8s.io/apimachinery/pkg/util/wait/wait.go:514 > k8s.io/kubernetes/test/e2e/network.testReachableUDP({0xc004c3de70, 0xe}, 0x7acb, 0x0?) test/e2e/network/service.go:603 > k8s.io/kubernetes/test/e2e/network.glob..func19.4() test/e2e/network/loadbalancer.go:390 k8s.io/kubernetes/vendor/github.com/onsi/ginkgo/v2/internal.extractBodyFunction.func3({0x2d591ce, 0xc0056e0000}) vendor/github.com/onsi/ginkgo/v2/internal/node.go:449 k8s.io/kubernetes/vendor/github.com/onsi/ginkgo/v2/internal.(*Suite).runNode.func2() vendor/github.com/onsi/ginkgo/v2/internal/suite.go:750 k8s.io/kubernetes/vendor/github.com/onsi/ginkgo/v2/internal.(*Suite).runNode vendor/github.com/onsi/ginkgo/v2/internal/suite.go:738 ------------------------------ Nov 26 05:41:01.120: INFO: Poking udp://35.247.104.236:31435 Nov 26 05:41:01.159: INFO: Poke("udp://35.247.104.236:31435"): read udp 10.60.29.184:53602->35.247.104.236:31435: read: connection refused Nov 26 05:41:03.120: INFO: Poking udp://35.247.104.236:31435 Nov 26 05:41:03.159: INFO: Poke("udp://35.247.104.236:31435"): read udp 10.60.29.184:57311->35.247.104.236:31435: read: connection refused Nov 26 05:41:05.119: INFO: Poking udp://35.247.104.236:31435 Nov 26 05:41:05.158: INFO: Poke("udp://35.247.104.236:31435"): read udp 10.60.29.184:35115->35.247.104.236:31435: read: connection refused Nov 26 05:41:07.120: INFO: Poking udp://35.247.104.236:31435 Nov 26 05:41:07.159: INFO: Poke("udp://35.247.104.236:31435"): read udp 10.60.29.184:52020->35.247.104.236:31435: read: connection refused Nov 26 05:41:09.122: INFO: Poking udp://35.247.104.236:31435 Nov 26 05:41:09.161: INFO: Poke("udp://35.247.104.236:31435"): read udp 10.60.29.184:48478->35.247.104.236:31435: read: connection refused Nov 26 05:41:11.119: INFO: Poking udp://35.247.104.236:31435 Nov 26 05:41:11.158: INFO: Poke("udp://35.247.104.236:31435"): read udp 10.60.29.184:57706->35.247.104.236:31435: read: connection refused Nov 26 05:41:13.120: INFO: Poking udp://35.247.104.236:31435 Nov 26 05:41:13.159: INFO: Poke("udp://35.247.104.236:31435"): read udp 10.60.29.184:44986->35.247.104.236:31435: read: connection refused Nov 26 05:41:15.120: INFO: Poking udp://35.247.104.236:31435 Nov 26 05:41:15.159: INFO: Poke("udp://35.247.104.236:31435"): read udp 10.60.29.184:46275->35.247.104.236:31435: read: connection refused Nov 26 05:41:17.120: INFO: Poking udp://35.247.104.236:31435 Nov 26 05:41:17.159: INFO: Poke("udp://35.247.104.236:31435"): read udp 10.60.29.184:45227->35.247.104.236:31435: read: connection refused Nov 26 05:41:19.119: INFO: Poking udp://35.247.104.236:31435 Nov 26 05:41:19.158: INFO: Poke("udp://35.247.104.236:31435"): read udp 10.60.29.184:42037->35.247.104.236:31435: read: connection refused ------------------------------ Progress Report for Ginkgo Process #6 Automatically polling progress: [sig-network] LoadBalancers should be able to change the type and ports of a UDP service [Slow] (Spec Runtime: 17m40.665s) test/e2e/network/loadbalancer.go:287 In [It] (Node Runtime: 17m40.103s) test/e2e/network/loadbalancer.go:287 At [By Step] hitting the UDP service's NodePort (Step Runtime: 3m13.656s) test/e2e/network/loadbalancer.go:389 Spec Goroutine goroutine 1768 [select] k8s.io/kubernetes/vendor/k8s.io/apimachinery/pkg/util/wait.WaitForWithContext({0x7fe0bc8, 0xc0000820c8}, 0xc00165d188, 0x2fdb16a?) vendor/k8s.io/apimachinery/pkg/util/wait/wait.go:660 k8s.io/kubernetes/vendor/k8s.io/apimachinery/pkg/util/wait.poll({0x7fe0bc8, 0xc0000820c8}, 0x60?, 0x2fd9d05?, 0x10?) vendor/k8s.io/apimachinery/pkg/util/wait/wait.go:596 k8s.io/kubernetes/vendor/k8s.io/apimachinery/pkg/util/wait.PollImmediateWithContext({0x7fe0bc8, 0xc0000820c8}, 0xc000202b80?, 0xc0046c5cb0?, 0x262a967?) vendor/k8s.io/apimachinery/pkg/util/wait/wait.go:528 k8s.io/kubernetes/vendor/k8s.io/apimachinery/pkg/util/wait.PollImmediate(0x65cbc00?, 0xc0002019c8?, 0x754e980?) vendor/k8s.io/apimachinery/pkg/util/wait/wait.go:514 > k8s.io/kubernetes/test/e2e/network.testReachableUDP({0xc004c3de70, 0xe}, 0x7acb, 0x0?) test/e2e/network/service.go:603 > k8s.io/kubernetes/test/e2e/network.glob..func19.4() test/e2e/network/loadbalancer.go:390 k8s.io/kubernetes/vendor/github.com/onsi/ginkgo/v2/internal.extractBodyFunction.func3({0x2d591ce, 0xc0056e0000}) vendor/github.com/onsi/ginkgo/v2/internal/node.go:449 k8s.io/kubernetes/vendor/github.com/onsi/ginkgo/v2/internal.(*Suite).runNode.func2() vendor/github.com/onsi/ginkgo/v2/internal/suite.go:750 k8s.io/kubernetes/vendor/github.com/onsi/ginkgo/v2/internal.(*Suite).runNode vendor/github.com/onsi/ginkgo/v2/internal/suite.go:738 ------------------------------ Nov 26 05:41:21.120: INFO: Poking udp://35.247.104.236:31435 Nov 26 05:41:21.159: INFO: Poke("udp://35.247.104.236:31435"): read udp 10.60.29.184:36845->35.247.104.236:31435: read: connection refused Nov 26 05:41:23.120: INFO: Poking udp://35.247.104.236:31435 Nov 26 05:41:23.159: INFO: Poke("udp://35.247.104.236:31435"): read udp 10.60.29.184:37857->35.247.104.236:31435: read: connection refused Nov 26 05:41:25.119: INFO: Poking udp://35.247.104.236:31435 Nov 26 05:41:25.159: INFO: Poke("udp://35.247.104.236:31435"): read udp 10.60.29.184:49454->35.247.104.236:31435: read: connection refused Nov 26 05:41:27.120: INFO: Poking udp://35.247.104.236:31435 Nov 26 05:41:27.159: INFO: Poke("udp://35.247.104.236:31435"): read udp 10.60.29.184:38036->35.247.104.236:31435: read: connection refused Nov 26 05:41:29.120: INFO: Poking udp://35.247.104.236:31435 Nov 26 05:41:29.159: INFO: Poke("udp://35.247.104.236:31435"): read udp 10.60.29.184:52869->35.247.104.236:31435: read: connection refused Nov 26 05:41:31.120: INFO: Poking udp://35.247.104.236:31435 Nov 26 05:41:31.159: INFO: Poke("udp://35.247.104.236:31435"): read udp 10.60.29.184:38943->35.247.104.236:31435: read: connection refused Nov 26 05:41:33.119: INFO: Poking udp://35.247.104.236:31435 Nov 26 05:41:33.158: INFO: Poke("udp://35.247.104.236:31435"): read udp 10.60.29.184:35921->35.247.104.236:31435: read: connection refused Nov 26 05:41:35.119: INFO: Poking udp://35.247.104.236:31435 Nov 26 05:41:35.158: INFO: Poke("udp://35.247.104.236:31435"): read udp 10.60.29.184:35809->35.247.104.236:31435: read: connection refused Nov 26 05:41:37.120: INFO: Poking udp://35.247.104.236:31435 Nov 26 05:41:37.159: INFO: Poke("udp://35.247.104.236:31435"): read udp 10.60.29.184:55819->35.247.104.236:31435: read: connection refused Nov 26 05:41:39.120: INFO: Poking udp://35.247.104.236:31435 Nov 26 05:41:39.159: INFO: Poke("udp://35.247.104.236:31435"): read udp 10.60.29.184:37606->35.247.104.236:31435: read: connection refused ------------------------------ Progress Report for Ginkgo Process #6 Automatically polling progress: [sig-network] LoadBalancers should be able to change the type and ports of a UDP service [Slow] (Spec Runtime: 18m0.668s) test/e2e/network/loadbalancer.go:287 In [It] (Node Runtime: 18m0.106s) test/e2e/network/loadbalancer.go:287 At [By Step] hitting the UDP service's NodePort (Step Runtime: 3m33.659s) test/e2e/network/loadbalancer.go:389 Spec Goroutine goroutine 1768 [select] k8s.io/kubernetes/vendor/k8s.io/apimachinery/pkg/util/wait.WaitForWithContext({0x7fe0bc8, 0xc0000820c8}, 0xc00165d188, 0x2fdb16a?) vendor/k8s.io/apimachinery/pkg/util/wait/wait.go:660 k8s.io/kubernetes/vendor/k8s.io/apimachinery/pkg/util/wait.poll({0x7fe0bc8, 0xc0000820c8}, 0x60?, 0x2fd9d05?, 0x10?) vendor/k8s.io/apimachinery/pkg/util/wait/wait.go:596 k8s.io/kubernetes/vendor/k8s.io/apimachinery/pkg/util/wait.PollImmediateWithContext({0x7fe0bc8, 0xc0000820c8}, 0xc000202b80?, 0xc0046c5cb0?, 0x262a967?) vendor/k8s.io/apimachinery/pkg/util/wait/wait.go:528 k8s.io/kubernetes/vendor/k8s.io/apimachinery/pkg/util/wait.PollImmediate(0x65cbc00?, 0xc0002019c8?, 0x754e980?) vendor/k8s.io/apimachinery/pkg/util/wait/wait.go:514 > k8s.io/kubernetes/test/e2e/network.testReachableUDP({0xc004c3de70, 0xe}, 0x7acb, 0x0?) test/e2e/network/service.go:603 > k8s.io/kubernetes/test/e2e/network.glob..func19.4() test/e2e/network/loadbalancer.go:390 k8s.io/kubernetes/vendor/github.com/onsi/ginkgo/v2/internal.extractBodyFunction.func3({0x2d591ce, 0xc0056e0000}) vendor/github.com/onsi/ginkgo/v2/internal/node.go:449 k8s.io/kubernetes/vendor/github.com/onsi/ginkgo/v2/internal.(*Suite).runNode.func2() vendor/github.com/onsi/ginkgo/v2/internal/suite.go:750 k8s.io/kubernetes/vendor/github.com/onsi/ginkgo/v2/internal.(*Suite).runNode vendor/github.com/onsi/ginkgo/v2/internal/suite.go:738 ------------------------------ Nov 26 05:41:41.120: INFO: Poking udp://35.247.104.236:31435 Nov 26 05:41:41.160: INFO: Poke("udp://35.247.104.236:31435"): read udp 10.60.29.184:59247->35.247.104.236:31435: read: connection refused Nov 26 05:41:43.119: INFO: Poking udp://35.247.104.236:31435 Nov 26 05:41:43.158: INFO: Poke("udp://35.247.104.236:31435"): read udp 10.60.29.184:48428->35.247.104.236:31435: read: connection refused Nov 26 05:41:45.119: INFO: Poking udp://35.247.104.236:31435 Nov 26 05:41:45.159: INFO: Poke("udp://35.247.104.236:31435"): read udp 10.60.29.184:52221->35.247.104.236:31435: read: connection refused Nov 26 05:41:47.120: INFO: Poking udp://35.247.104.236:31435 Nov 26 05:41:47.159: INFO: Poke("udp://35.247.104.236:31435"): read udp 10.60.29.184:35161->35.247.104.236:31435: read: connection refused Nov 26 05:41:49.120: INFO: Poking udp://35.247.104.236:31435 Nov 26 05:41:49.159: INFO: Poke("udp://35.247.104.236:31435"): read udp 10.60.29.184:58711->35.247.104.236:31435: read: connection refused Nov 26 05:41:51.120: INFO: Poking udp://35.247.104.236:31435 Nov 26 05:41:51.159: INFO: Poke("udp://35.247.104.236:31435"): read udp 10.60.29.184:36662->35.247.104.236:31435: read: connection refused Nov 26 05:41:53.120: INFO: Poking udp://35.247.104.236:31435 Nov 26 05:41:53.159: INFO: Poke("udp://35.247.104.236:31435"): read udp 10.60.29.184:47307->35.247.104.236:31435: read: connection refused Nov 26 05:41:55.119: INFO: Poking udp://35.247.104.236:31435 Nov 26 05:41:55.159: INFO: Poke("udp://35.247.104.236:31435"): read udp 10.60.29.184:47483->35.247.104.236:31435: read: connection refused Nov 26 05:41:57.120: INFO: Poking udp://35.247.104.236:31435 Nov 26 05:41:57.159: INFO: Poke("udp://35.247.104.236:31435"): read udp 10.60.29.184:52835->35.247.104.236:31435: read: connection refused Nov 26 05:41:59.120: INFO: Poking udp://35.247.104.236:31435 Nov 26 05:41:59.159: INFO: Poke("udp://35.247.104.236:31435"): read udp 10.60.29.184:40387->35.247.104.236:31435: read: connection refused ------------------------------ Progress Report for Ginkgo Process #6 Automatically polling progress: [sig-network] LoadBalancers should be able to change the type and ports of a UDP service [Slow] (Spec Runtime: 18m20.67s) test/e2e/network/loadbalancer.go:287 In [It] (Node Runtime: 18m20.107s) test/e2e/network/loadbalancer.go:287 At [By Step] hitting the UDP service's NodePort (Step Runtime: 3m53.661s) test/e2e/network/loadbalancer.go:389 Spec Goroutine goroutine 1768 [select] k8s.io/kubernetes/vendor/k8s.io/apimachinery/pkg/util/wait.WaitForWithContext({0x7fe0bc8, 0xc0000820c8}, 0xc00165d188, 0x2fdb16a?) vendor/k8s.io/apimachinery/pkg/util/wait/wait.go:660 k8s.io/kubernetes/vendor/k8s.io/apimachinery/pkg/util/wait.poll({0x7fe0bc8, 0xc0000820c8}, 0x60?, 0x2fd9d05?, 0x10?) vendor/k8s.io/apimachinery/pkg/util/wait/wait.go:596 k8s.io/kubernetes/vendor/k8s.io/apimachinery/pkg/util/wait.PollImmediateWithContext({0x7fe0bc8, 0xc0000820c8}, 0xc000202b80?, 0xc0046c5cb0?, 0x262a967?) vendor/k8s.io/apimachinery/pkg/util/wait/wait.go:528 k8s.io/kubernetes/vendor/k8s.io/apimachinery/pkg/util/wait.PollImmediate(0x65cbc00?, 0xc0002019c8?, 0x754e980?) vendor/k8s.io/apimachinery/pkg/util/wait/wait.go:514 > k8s.io/kubernetes/test/e2e/network.testReachableUDP({0xc004c3de70, 0xe}, 0x7acb, 0x0?) test/e2e/network/service.go:603 > k8s.io/kubernetes/test/e2e/network.glob..func19.4() test/e2e/network/loadbalancer.go:390 k8s.io/kubernetes/vendor/github.com/onsi/ginkgo/v2/internal.extractBodyFunction.func3({0x2d591ce, 0xc0056e0000}) vendor/github.com/onsi/ginkgo/v2/internal/node.go:449 k8s.io/kubernetes/vendor/github.com/onsi/ginkgo/v2/internal.(*Suite).runNode.func2() vendor/github.com/onsi/ginkgo/v2/internal/suite.go:750 k8s.io/kubernetes/vendor/github.com/onsi/ginkgo/v2/internal.(*Suite).runNode vendor/github.com/onsi/ginkgo/v2/internal/suite.go:738 ------------------------------ Nov 26 05:42:01.120: INFO: Poking udp://35.247.104.236:31435 Nov 26 05:42:01.159: INFO: Poke("udp://35.247.104.236:31435"): read udp 10.60.29.184:42340->35.247.104.236:31435: read: connection refused Nov 26 05:42:03.120: INFO: Poking udp://35.247.104.236:31435 Nov 26 05:42:03.159: INFO: Poke("udp://35.247.104.236:31435"): read udp 10.60.29.184:49781->35.247.104.236:31435: read: connection refused Nov 26 05:42:05.119: INFO: Poking udp://35.247.104.236:31435 Nov 26 05:42:05.160: INFO: Poke("udp://35.247.104.236:31435"): read udp 10.60.29.184:48248->35.247.104.236:31435: read: connection refused Nov 26 05:42:07.120: INFO: Poking udp://35.247.104.236:31435 Nov 26 05:42:07.159: INFO: Poke("udp://35.247.104.236:31435"): read udp 10.60.29.184:43040->35.247.104.236:31435: read: connection refused Nov 26 05:42:09.120: INFO: Poking udp://35.247.104.236:31435 Nov 26 05:42:09.159: INFO: Poke("udp://35.247.104.236:31435"): read udp 10.60.29.184:41446->35.247.104.236:31435: read: connection refused Nov 26 05:42:11.120: INFO: Poking udp://35.247.104.236:31435 Nov 26 05:42:11.159: INFO: Poke("udp://35.247.104.236:31435"): read udp 10.60.29.184:40437->35.247.104.236:31435: read: connection refused Nov 26 05:42:13.120: INFO: Poking udp://35.247.104.236:31435 Nov 26 05:42:13.159: INFO: Poke("udp://35.247.104.236:31435"): read udp 10.60.29.184:58845->35.247.104.236:31435: read: connection refused Nov 26 05:42:15.120: INFO: Poking udp://35.247.104.236:31435 Nov 26 05:42:15.159: INFO: Poke("udp://35.247.104.236:31435"): read udp 10.60.29.184:35338->35.247.104.236:31435: read: connection refused Nov 26 05:42:17.120: INFO: Poking udp://35.247.104.236:31435 Nov 26 05:42:17.159: INFO: Poke("udp://35.247.104.236:31435"): read udp 10.60.29.184:36240->35.247.104.236:31435: read: connection refused Nov 26 05:42:19.120: INFO: Poking udp://35.247.104.236:31435 Nov 26 05:42:19.159: INFO: Poke("udp://35.247.104.236:31435"): read udp 10.60.29.184:41337->35.247.104.236:31435: read: connection refused ------------------------------ Progress Report for Ginkgo Process #6 Automatically polling progress: [sig-network] LoadBalancers should be able to change the type and ports of a UDP service [Slow] (Spec Runtime: 18m40.673s) test/e2e/network/loadbalancer.go:287 In [It] (Node Runtime: 18m40.111s) test/e2e/network/loadbalancer.go:287 At [By Step] hitting the UDP service's NodePort (Step Runtime: 4m13.664s) test/e2e/network/loadbalancer.go:389 Spec Goroutine goroutine 1768 [select] k8s.io/kubernetes/vendor/k8s.io/apimachinery/pkg/util/wait.WaitForWithContext({0x7fe0bc8, 0xc0000820c8}, 0xc00165d188, 0x2fdb16a?) vendor/k8s.io/apimachinery/pkg/util/wait/wait.go:660 k8s.io/kubernetes/vendor/k8s.io/apimachinery/pkg/util/wait.poll({0x7fe0bc8, 0xc0000820c8}, 0x60?, 0x2fd9d05?, 0x10?) vendor/k8s.io/apimachinery/pkg/util/wait/wait.go:596 k8s.io/kubernetes/vendor/k8s.io/apimachinery/pkg/util/wait.PollImmediateWithContext({0x7fe0bc8, 0xc0000820c8}, 0xc000202b80?, 0xc0046c5cb0?, 0x262a967?) vendor/k8s.io/apimachinery/pkg/util/wait/wait.go:528 k8s.io/kubernetes/vendor/k8s.io/apimachinery/pkg/util/wait.PollImmediate(0x65cbc00?, 0xc0002019c8?, 0x754e980?) vendor/k8s.io/apimachinery/pkg/util/wait/wait.go:514 > k8s.io/kubernetes/test/e2e/network.testReachableUDP({0xc004c3de70, 0xe}, 0x7acb, 0x0?) test/e2e/network/service.go:603 > k8s.io/kubernetes/test/e2e/network.glob..func19.4() test/e2e/network/loadbalancer.go:390 k8s.io/kubernetes/vendor/github.com/onsi/ginkgo/v2/internal.extractBodyFunction.func3({0x2d591ce, 0xc0056e0000}) vendor/github.com/onsi/ginkgo/v2/internal/node.go:449 k8s.io/kubernetes/vendor/github.com/onsi/ginkgo/v2/internal.(*Suite).runNode.func2() vendor/github.com/onsi/ginkgo/v2/internal/suite.go:750 k8s.io/kubernetes/vendor/github.com/onsi/ginkgo/v2/internal.(*Suite).runNode vendor/github.com/onsi/ginkgo/v2/internal/suite.go:738 ------------------------------ Nov 26 05:42:21.119: INFO: Poking udp://35.247.104.236:31435 Nov 26 05:42:21.158: INFO: Poke("udp://35.247.104.236:31435"): read udp 10.60.29.184:60500->35.247.104.236:31435: read: connection refused Nov 26 05:42:23.120: INFO: Poking udp://35.247.104.236:31435 Nov 26 05:42:23.159: INFO: Poke("udp://35.247.104.236:31435"): read udp 10.60.29.184:41414->35.247.104.236:31435: read: connection refused Nov 26 05:42:25.120: INFO: Poking udp://35.247.104.236:31435 Nov 26 05:42:25.159: INFO: Poke("udp://35.247.104.236:31435"): read udp 10.60.29.184:33790->35.247.104.236:31435: read: connection refused Nov 26 05:42:27.120: INFO: Poking udp://35.247.104.236:31435 Nov 26 05:42:27.159: INFO: Poke("udp://35.247.104.236:31435"): read udp 10.60.29.184:33489->35.247.104.236:31435: read: connection refused Nov 26 05:42:29.119: INFO: Poking udp://35.247.104.236:31435 Nov 26 05:42:29.158: INFO: Poke("udp://35.247.104.236:31435"): read udp 10.60.29.184:47740->35.247.104.236:31435: read: connection refused Nov 26 05:42:31.120: INFO: Poking udp://35.247.104.236:31435 Nov 26 05:42:31.159: INFO: Poke("udp://35.247.104.236:31435"): read udp 10.60.29.184:36730->35.247.104.236:31435: read: connection refused Nov 26 05:42:33.119: INFO: Poking udp://35.247.104.236:31435 Nov 26 05:42:33.159: INFO: Poke("udp://35.247.104.236:31435"): read udp 10.60.29.184:57428->35.247.104.236:31435: read: connection refused Nov 26 05:42:35.120: INFO: Poking udp://35.247.104.236:31435 Nov 26 05:42:35.159: INFO: Poke("udp://35.247.104.236:31435"): read udp 10.60.29.184:55367->35.247.104.236:31435: read: connection refused Nov 26 05:42:37.120: INFO: Poking udp://35.247.104.236:31435 Nov 26 05:42:37.160: INFO: Poke("udp://35.247.104.236:31435"): read udp 10.60.29.184:59465->35.247.104.236:31435: read: connection refused Nov 26 05:42:39.120: INFO: Poking udp://35.247.104.236:31435 Nov 26 05:42:39.159: INFO: Poke("udp://35.247.104.236:31435"): read udp 10.60.29.184:33279->35.247.104.236:31435: read: connection refused ------------------------------ Progress Report for Ginkgo Process #6 Automatically polling progress: [sig-network] LoadBalancers should be able to change the type and ports of a UDP service [Slow] (Spec Runtime: 19m0.674s) test/e2e/network/loadbalancer.go:287 In [It] (Node Runtime: 19m0.112s) test/e2e/network/loadbalancer.go:287 At [By Step] hitting the UDP service's NodePort (Step Runtime: 4m33.665s) test/e2e/network/loadbalancer.go:389 Spec Goroutine goroutine 1768 [select] k8s.io/kubernetes/vendor/k8s.io/apimachinery/pkg/util/wait.WaitForWithContext({0x7fe0bc8, 0xc0000820c8}, 0xc00165d188, 0x2fdb16a?) vendor/k8s.io/apimachinery/pkg/util/wait/wait.go:660 k8s.io/kubernetes/vendor/k8s.io/apimachinery/pkg/util/wait.poll({0x7fe0bc8, 0xc0000820c8}, 0x60?, 0x2fd9d05?, 0x10?) vendor/k8s.io/apimachinery/pkg/util/wait/wait.go:596 k8s.io/kubernetes/vendor/k8s.io/apimachinery/pkg/util/wait.PollImmediateWithContext({0x7fe0bc8, 0xc0000820c8}, 0xc000202b80?, 0xc0046c5cb0?, 0x262a967?) vendor/k8s.io/apimachinery/pkg/util/wait/wait.go:528 k8s.io/kubernetes/vendor/k8s.io/apimachinery/pkg/util/wait.PollImmediate(0x65cbc00?, 0xc0002019c8?, 0x754e980?) vendor/k8s.io/apimachinery/pkg/util/wait/wait.go:514 > k8s.io/kubernetes/test/e2e/network.testReachableUDP({0xc004c3de70, 0xe}, 0x7acb, 0x0?) test/e2e/network/service.go:603 > k8s.io/kubernetes/test/e2e/network.glob..func19.4() test/e2e/network/loadbalancer.go:390 k8s.io/kubernetes/vendor/github.com/onsi/ginkgo/v2/internal.extractBodyFunction.func3({0x2d591ce, 0xc0056e0000}) vendor/github.com/onsi/ginkgo/v2/internal/node.go:449 k8s.io/kubernetes/vendor/github.com/onsi/ginkgo/v2/internal.(*Suite).runNode.func2() vendor/github.com/onsi/ginkgo/v2/internal/suite.go:750 k8s.io/kubernetes/vendor/github.com/onsi/ginkgo/v2/internal.(*Suite).runNode vendor/github.com/onsi/ginkgo/v2/internal/suite.go:738 ------------------------------ Nov 26 05:42:41.120: INFO: Poking udp://35.247.104.236:31435 Nov 26 05:42:41.159: INFO: Poke("udp://35.247.104.236:31435"): read udp 10.60.29.184:37466->35.247.104.236:31435: read: connection refused Nov 26 05:42:43.120: INFO: Poking udp://35.247.104.236:31435 Nov 26 05:42:43.159: INFO: Poke("udp://35.247.104.236:31435"): read udp 10.60.29.184:38945->35.247.104.236:31435: read: connection refused Nov 26 05:42:45.119: INFO: Poking udp://35.247.104.236:31435 Nov 26 05:42:45.159: INFO: Poke("udp://35.247.104.236:31435"): read udp 10.60.29.184:53872->35.247.104.236:31435: read: connection refused Nov 26 05:42:47.119: INFO: Poking udp://35.247.104.236:31435 Nov 26 05:42:47.159: INFO: Poke("udp://35.247.104.236:31435"): read udp 10.60.29.184:33153->35.247.104.236:31435: read: connection refused Nov 26 05:42:49.120: INFO: Poking udp://35.247.104.236:31435 Nov 26 05:42:49.159: INFO: Poke("udp://35.247.104.236:31435"): read udp 10.60.29.184:54762->35.247.104.236:31435: read: connection refused Nov 26 05:42:51.120: INFO: Poking udp://35.247.104.236:31435 Nov 26 05:42:51.159: INFO: Poke("udp://35.247.104.236:31435"): read udp 10.60.29.184:47590->35.247.104.236:31435: read: connection refused Nov 26 05:42:53.119: INFO: Poking udp://35.247.104.236:31435 Nov 26 05:42:53.158: INFO: Poke("udp://35.247.104.236:31435"): read udp 10.60.29.184:32836->35.247.104.236:31435: read: connection refused Nov 26 05:42:55.119: INFO: Poking udp://35.247.104.236:31435 Nov 26 05:42:55.158: INFO: Poke("udp://35.247.104.236:31435"): read udp 10.60.29.184:34555->35.247.104.236:31435: read: connection refused Nov 26 05:42:57.120: INFO: Poking udp://35.247.104.236:31435 Nov 26 05:42:57.159: INFO: Poke("udp://35.247.104.236:31435"): read udp 10.60.29.184:33070->35.247.104.236:31435: read: connection refused Nov 26 05:42:59.120: INFO: Poking udp://35.247.104.236:31435 Nov 26 05:42:59.159: INFO: Poke("udp://35.247.104.236:31435"): read udp 10.60.29.184:41005->35.247.104.236:31435: read: connection refused ------------------------------ Progress Report for Ginkgo Process #6 Automatically polling progress: [sig-network] LoadBalancers should be able to change the type and ports of a UDP service [Slow] (Spec Runtime: 19m20.676s) test/e2e/network/loadbalancer.go:287 In [It] (Node Runtime: 19m20.113s) test/e2e/network/loadbalancer.go:287 At [By Step] hitting the UDP service's NodePort (Step Runtime: 4m53.666s) test/e2e/network/loadbalancer.go:389 Spec Goroutine goroutine 1768 [select] k8s.io/kubernetes/vendor/k8s.io/apimachinery/pkg/util/wait.WaitForWithContext({0x7fe0bc8, 0xc0000820c8}, 0xc00165d188, 0x2fdb16a?) vendor/k8s.io/apimachinery/pkg/util/wait/wait.go:660 k8s.io/kubernetes/vendor/k8s.io/apimachinery/pkg/util/wait.poll({0x7fe0bc8, 0xc0000820c8}, 0x60?, 0x2fd9d05?, 0x10?) vendor/k8s.io/apimachinery/pkg/util/wait/wait.go:596 k8s.io/kubernetes/vendor/k8s.io/apimachinery/pkg/util/wait.PollImmediateWithContext({0x7fe0bc8, 0xc0000820c8}, 0xc000202b80?, 0xc0046c5cb0?, 0x262a967?) vendor/k8s.io/apimachinery/pkg/util/wait/wait.go:528 k8s.io/kubernetes/vendor/k8s.io/apimachinery/pkg/util/wait.PollImmediate(0x65cbc00?, 0xc0002019c8?, 0x754e980?) vendor/k8s.io/apimachinery/pkg/util/wait/wait.go:514 > k8s.io/kubernetes/test/e2e/network.testReachableUDP({0xc004c3de70, 0xe}, 0x7acb, 0x0?) test/e2e/network/service.go:603 > k8s.io/kubernetes/test/e2e/network.glob..func19.4() test/e2e/network/loadbalancer.go:390 k8s.io/kubernetes/vendor/github.com/onsi/ginkgo/v2/internal.extractBodyFunction.func3({0x2d591ce, 0xc0056e0000}) vendor/github.com/onsi/ginkgo/v2/internal/node.go:449 k8s.io/kubernetes/vendor/github.com/onsi/ginkgo/v2/internal.(*Suite).runNode.func2() vendor/github.com/onsi/ginkgo/v2/internal/suite.go:750 k8s.io/kubernetes/vendor/github.com/onsi/ginkgo/v2/internal.(*Suite).runNode vendor/github.com/onsi/ginkgo/v2/internal/suite.go:738 ------------------------------ Nov 26 05:43:01.120: INFO: Poking udp://35.247.104.236:31435 Nov 26 05:43:01.159: INFO: Poke("udp://35.247.104.236:31435"): read udp 10.60.29.184:52694->35.247.104.236:31435: read: connection refused Nov 26 05:43:03.119: INFO: Poking udp://35.247.104.236:31435 Nov 26 05:43:03.158: INFO: Poke("udp://35.247.104.236:31435"): read udp 10.60.29.184:51704->35.247.104.236:31435: read: connection refused Nov 26 05:43:05.120: INFO: Poking udp://35.247.104.236:31435 Nov 26 05:43:05.165: INFO: Poke("udp://35.247.104.236:31435"): read udp 10.60.29.184:47935->35.247.104.236:31435: read: connection refused Nov 26 05:43:07.120: INFO: Poking udp://35.247.104.236:31435 Nov 26 05:43:07.159: INFO: Poke("udp://35.247.104.236:31435"): read udp 10.60.29.184:42190->35.247.104.236:31435: read: connection refused Nov 26 05:43:07.159: INFO: Poking udp://35.247.104.236:31435 Nov 26 05:43:07.198: INFO: Poke("udp://35.247.104.236:31435"): read udp 10.60.29.184:50306->35.247.104.236:31435: read: connection refused Nov 26 05:43:07.198: FAIL: Could not reach UDP service through 35.247.104.236:31435 after 5m0s: timed out waiting for the condition Full Stack Trace k8s.io/kubernetes/test/e2e/network.testReachableUDP({0xc004c3de70, 0xe}, 0x7acb, 0x0?) test/e2e/network/service.go:604 +0x17b k8s.io/kubernetes/test/e2e/network.glob..func19.4() test/e2e/network/loadbalancer.go:390 +0xb25 [AfterEach] [sig-network] LoadBalancers test/e2e/framework/node/init/init.go:32 Nov 26 05:43:07.199: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready [AfterEach] [sig-network] LoadBalancers test/e2e/network/loadbalancer.go:71 Nov 26 05:43:07.366: INFO: Output of kubectl describe svc: Nov 26 05:43:07.366: INFO: Running '/workspace/github.com/containerd/containerd/kubernetes/platforms/linux/amd64/kubectl --server=https://35.227.179.144 --kubeconfig=/workspace/.kube/config --namespace=loadbalancers-6791 describe svc --namespace=loadbalancers-6791' Nov 26 05:43:07.695: INFO: stderr: "" Nov 26 05:43:07.695: INFO: stdout: "Name: mutability-test\nNamespace: loadbalancers-6791\nLabels: testid=mutability-test-80bd5d18-eb3d-4a05-8d82-e8975011c4f9\nAnnotations: <none>\nSelector: testid=mutability-test-80bd5d18-eb3d-4a05-8d82-e8975011c4f9\nType: LoadBalancer\nIP Family Policy: SingleStack\nIP Families: IPv4\nIP: 10.0.168.0\nIPs: 10.0.168.0\nLoadBalancer Ingress: 35.247.7.127\nPort: <unset> 80/UDP\nTargetPort: 80/UDP\nNodePort: <unset> 31435/UDP\nEndpoints: \nSession Affinity: None\nExternal Traffic Policy: Cluster\nEvents:\n Type Reason Age From Message\n ---- ------ ---- ---- -------\n Normal EnsuringLoadBalancer 5m29s service-controller Ensuring load balancer\n Normal EnsuredLoadBalancer 5m1s service-controller Ensured load balancer\n" Nov 26 05:43:07.695: INFO: Name: mutability-test Namespace: loadbalancers-6791 Labels: testid=mutability-test-80bd5d18-eb3d-4a05-8d82-e8975011c4f9 Annotations: <none> Selector: testid=mutability-test-80bd5d18-eb3d-4a05-8d82-e8975011c4f9 Type: LoadBalancer IP Family Policy: SingleStack IP Families: IPv4 IP: 10.0.168.0 IPs: 10.0.168.0 LoadBalancer Ingress: 35.247.7.127 Port: <unset> 80/UDP TargetPort: 80/UDP NodePort: <unset> 31435/UDP Endpoints: Session Affinity: None External Traffic Policy: Cluster Events: Type Reason Age From Message ---- ------ ---- ---- ------- Normal EnsuringLoadBalancer 5m29s service-controller Ensuring load balancer Normal EnsuredLoadBalancer 5m1s service-controller Ensured load balancer [DeferCleanup (Each)] [sig-network] LoadBalancers test/e2e/framework/metrics/init/init.go:33 [DeferCleanup (Each)] [sig-network] LoadBalancers dump namespaces | framework.go:196 STEP: dump namespace information after failure 11/26/22 05:43:07.696 STEP: Collecting events from namespace "loadbalancers-6791". 11/26/22 05:43:07.696 STEP: Found 11 events. 11/26/22 05:43:07.738 Nov 26 05:43:07.738: INFO: At 0001-01-01 00:00:00 +0000 UTC - event for mutability-test-7ktz4: { } Scheduled: Successfully assigned loadbalancers-6791/mutability-test-7ktz4 to bootstrap-e2e-minion-group-vhdj Nov 26 05:43:07.738: INFO: At 2022-11-26 05:23:40 +0000 UTC - event for mutability-test: {replication-controller } SuccessfulCreate: Created pod: mutability-test-7ktz4 Nov 26 05:43:07.738: INFO: At 2022-11-26 05:23:42 +0000 UTC - event for mutability-test-7ktz4: {kubelet bootstrap-e2e-minion-group-vhdj} Pulled: Container image "registry.k8s.io/e2e-test-images/agnhost:2.43" already present on machine Nov 26 05:43:07.738: INFO: At 2022-11-26 05:23:42 +0000 UTC - event for mutability-test-7ktz4: {kubelet bootstrap-e2e-minion-group-vhdj} Created: Created container netexec Nov 26 05:43:07.738: INFO: At 2022-11-26 05:23:42 +0000 UTC - event for mutability-test-7ktz4: {kubelet bootstrap-e2e-minion-group-vhdj} Started: Started container netexec Nov 26 05:43:07.738: INFO: At 2022-11-26 05:26:24 +0000 UTC - event for mutability-test-7ktz4: {kubelet bootstrap-e2e-minion-group-vhdj} Killing: Stopping container netexec Nov 26 05:43:07.738: INFO: At 2022-11-26 05:26:25 +0000 UTC - event for mutability-test-7ktz4: {kubelet bootstrap-e2e-minion-group-vhdj} SandboxChanged: Pod sandbox changed, it will be killed and re-created. Nov 26 05:43:07.738: INFO: At 2022-11-26 05:26:27 +0000 UTC - event for mutability-test-7ktz4: {kubelet bootstrap-e2e-minion-group-vhdj} BackOff: Back-off restarting failed container netexec in pod mutability-test-7ktz4_loadbalancers-6791(5ba9e818-4545-45fb-a290-03f35164ec45) Nov 26 05:43:07.738: INFO: At 2022-11-26 05:26:27 +0000 UTC - event for mutability-test-7ktz4: {kubelet bootstrap-e2e-minion-group-vhdj} Unhealthy: Readiness probe failed: Get "http://10.64.3.182:80/hostName": context deadline exceeded (Client.Timeout exceeded while awaiting headers) Nov 26 05:43:07.738: INFO: At 2022-11-26 05:37:38 +0000 UTC - event for mutability-test: {service-controller } EnsuringLoadBalancer: Ensuring load balancer Nov 26 05:43:07.738: INFO: At 2022-11-26 05:38:06 +0000 UTC - event for mutability-test: {service-controller } EnsuredLoadBalancer: Ensured load balancer Nov 26 05:43:07.779: INFO: POD NODE PHASE GRACE CONDITIONS Nov 26 05:43:07.779: INFO: mutability-test-7ktz4 bootstrap-e2e-minion-group-vhdj Running [{Initialized True 0001-01-01 00:00:00 +0000 UTC 2022-11-26 05:23:40 +0000 UTC } {Ready False 0001-01-01 00:00:00 +0000 UTC 2022-11-26 05:39:54 +0000 UTC ContainersNotReady containers with unready status: [netexec]} {ContainersReady False 0001-01-01 00:00:00 +0000 UTC 2022-11-26 05:39:54 +0000 UTC ContainersNotReady containers with unready status: [netexec]} {PodScheduled True 0001-01-01 00:00:00 +0000 UTC 2022-11-26 05:23:40 +0000 UTC }] Nov 26 05:43:07.779: INFO: Nov 26 05:43:07.823: INFO: Unable to fetch loadbalancers-6791/mutability-test-7ktz4/netexec logs: an error on the server ("unknown") has prevented the request from succeeding (get pods mutability-test-7ktz4) Nov 26 05:43:07.871: INFO: Logging node info for node bootstrap-e2e-master Nov 26 05:43:07.912: INFO: Node Info: &Node{ObjectMeta:{bootstrap-e2e-master 9a48eb11-c964-4bf8-82c0-1c32a9703981 13849 0 2022-11-26 05:05:18 +0000 UTC <nil> <nil> map[beta.kubernetes.io/arch:amd64 beta.kubernetes.io/instance-type:n1-standard-1 beta.kubernetes.io/os:linux cloud.google.com/metadata-proxy-ready:true failure-domain.beta.kubernetes.io/region:us-west1 failure-domain.beta.kubernetes.io/zone:us-west1-b kubernetes.io/arch:amd64 kubernetes.io/hostname:bootstrap-e2e-master kubernetes.io/os:linux node.kubernetes.io/instance-type:n1-standard-1 topology.kubernetes.io/region:us-west1 topology.kubernetes.io/zone:us-west1-b] map[node.alpha.kubernetes.io/ttl:0 volumes.kubernetes.io/controller-managed-attach-detach:true] [] [] [{kubelet Update v1 2022-11-26 05:05:18 +0000 UTC FieldsV1 {"f:metadata":{"f:annotations":{".":{},"f:volumes.kubernetes.io/controller-managed-attach-detach":{}},"f:labels":{".":{},"f:beta.kubernetes.io/arch":{},"f:beta.kubernetes.io/instance-type":{},"f:beta.kubernetes.io/os":{},"f:cloud.google.com/metadata-proxy-ready":{},"f:failure-domain.beta.kubernetes.io/region":{},"f:failure-domain.beta.kubernetes.io/zone":{},"f:kubernetes.io/arch":{},"f:kubernetes.io/hostname":{},"f:kubernetes.io/os":{},"f:node.kubernetes.io/instance-type":{},"f:topology.kubernetes.io/region":{},"f:topology.kubernetes.io/zone":{}}},"f:spec":{"f:providerID":{},"f:unschedulable":{}}} } {kube-controller-manager Update v1 2022-11-26 05:05:25 +0000 UTC FieldsV1 {"f:status":{"f:conditions":{"k:{\"type\":\"NetworkUnavailable\"}":{"f:lastHeartbeatTime":{},"f:lastTransitionTime":{},"f:message":{},"f:reason":{},"f:status":{}}}}} status} {kube-controller-manager Update v1 2022-11-26 05:05:39 +0000 UTC FieldsV1 {"f:metadata":{"f:annotations":{"f:node.alpha.kubernetes.io/ttl":{}}},"f:spec":{"f:podCIDR":{},"f:podCIDRs":{".":{},"v:\"10.64.1.0/24\"":{}},"f:taints":{}}} } {kubelet Update v1 2022-11-26 05:42:56 +0000 UTC FieldsV1 {"f:status":{"f:conditions":{"k:{\"type\":\"DiskPressure\"}":{"f:lastHeartbeatTime":{}},"k:{\"type\":\"MemoryPressure\"}":{"f:lastHeartbeatTime":{}},"k:{\"type\":\"PIDPressure\"}":{"f:lastHeartbeatTime":{}},"k:{\"type\":\"Ready\"}":{"f:lastHeartbeatTime":{},"f:lastTransitionTime":{},"f:message":{},"f:reason":{},"f:status":{}}},"f:images":{}}} status}]},Spec:NodeSpec{PodCIDR:10.64.1.0/24,DoNotUseExternalID:,ProviderID:gce://k8s-jkns-e2e-gce-ci-reboot/us-west1-b/bootstrap-e2e-master,Unschedulable:true,Taints:[]Taint{Taint{Key:node-role.kubernetes.io/master,Value:,Effect:NoSchedule,TimeAdded:<nil>,},Taint{Key:node.kubernetes.io/unschedulable,Value:,Effect:NoSchedule,TimeAdded:<nil>,},},ConfigSource:nil,PodCIDRs:[10.64.1.0/24],},Status:NodeStatus{Capacity:ResourceList{attachable-volumes-gce-pd: {{127 0} {<nil>} 127 DecimalSI},cpu: {{1 0} {<nil>} 1 DecimalSI},ephemeral-storage: {{16656896000 0} {<nil>} 16266500Ki BinarySI},hugepages-1Gi: {{0 0} {<nil>} 0 DecimalSI},hugepages-2Mi: {{0 0} {<nil>} 0 DecimalSI},memory: {{3858374656 0} {<nil>} BinarySI},pods: {{110 0} {<nil>} 110 DecimalSI},},Allocatable:ResourceList{attachable-volumes-gce-pd: {{127 0} {<nil>} 127 DecimalSI},cpu: {{1 0} {<nil>} 1 DecimalSI},ephemeral-storage: {{14991206376 0} {<nil>} 14991206376 DecimalSI},hugepages-1Gi: {{0 0} {<nil>} 0 DecimalSI},hugepages-2Mi: {{0 0} {<nil>} 0 DecimalSI},memory: {{3596230656 0} {<nil>} BinarySI},pods: {{110 0} {<nil>} 110 DecimalSI},},Phase:,Conditions:[]NodeCondition{NodeCondition{Type:NetworkUnavailable,Status:False,LastHeartbeatTime:2022-11-26 05:05:25 +0000 UTC,LastTransitionTime:2022-11-26 05:05:25 +0000 UTC,Reason:RouteCreated,Message:RouteController created a route,},NodeCondition{Type:MemoryPressure,Status:False,LastHeartbeatTime:2022-11-26 05:42:56 +0000 UTC,LastTransitionTime:2022-11-26 05:05:18 +0000 UTC,Reason:KubeletHasSufficientMemory,Message:kubelet has sufficient memory available,},NodeCondition{Type:DiskPressure,Status:False,LastHeartbeatTime:2022-11-26 05:42:56 +0000 UTC,LastTransitionTime:2022-11-26 05:05:18 +0000 UTC,Reason:KubeletHasNoDiskPressure,Message:kubelet has no disk pressure,},NodeCondition{Type:PIDPressure,Status:False,LastHeartbeatTime:2022-11-26 05:42:56 +0000 UTC,LastTransitionTime:2022-11-26 05:05:18 +0000 UTC,Reason:KubeletHasSufficientPID,Message:kubelet has sufficient PID available,},NodeCondition{Type:Ready,Status:True,LastHeartbeatTime:2022-11-26 05:42:56 +0000 UTC,LastTransitionTime:2022-11-26 05:05:38 +0000 UTC,Reason:KubeletReady,Message:kubelet is posting ready status. AppArmor enabled,},},Addresses:[]NodeAddress{NodeAddress{Type:InternalIP,Address:10.138.0.2,},NodeAddress{Type:ExternalIP,Address:35.227.179.144,},NodeAddress{Type:InternalDNS,Address:bootstrap-e2e-master.c.k8s-jkns-e2e-gce-ci-reboot.internal,},NodeAddress{Type:Hostname,Address:bootstrap-e2e-master.c.k8s-jkns-e2e-gce-ci-reboot.internal,},},DaemonEndpoints:NodeDaemonEndpoints{KubeletEndpoint:DaemonEndpoint{Port:10250,},},NodeInfo:NodeSystemInfo{MachineID:0fdb3cfe29f66637553465718381a2f8,SystemUUID:0fdb3cfe-29f6-6637-5534-65718381a2f8,BootID:fb804a29-7e61-434d-8044-6944159ce35f,KernelVersion:5.10.123+,OSImage:Container-Optimized OS from Google,ContainerRuntimeVersion:containerd://1.7.0-beta.0-149-gd06318622,KubeletVersion:v1.27.0-alpha.0.50+70617042976dc1,KubeProxyVersion:v1.27.0-alpha.0.50+70617042976dc1,OperatingSystem:linux,Architecture:amd64,},Images:[]ContainerImage{ContainerImage{Names:[registry.k8s.io/kube-apiserver-amd64:v1.27.0-alpha.0.50_70617042976dc1],SizeBytes:135160272,},ContainerImage{Names:[registry.k8s.io/kube-controller-manager-amd64:v1.27.0-alpha.0.50_70617042976dc1],SizeBytes:124990265,},ContainerImage{Names:[registry.k8s.io/etcd@sha256:dd75ec974b0a2a6f6bb47001ba09207976e625db898d1b16735528c009cb171c registry.k8s.io/etcd:3.5.6-0],SizeBytes:102542580,},ContainerImage{Names:[registry.k8s.io/kube-scheduler-amd64:v1.27.0-alpha.0.50_70617042976dc1],SizeBytes:57660216,},ContainerImage{Names:[gke.gcr.io/prometheus-to-sd@sha256:e739643c3939ba0b161425f45a1989eedfc4a3b166db9a7100863296b4c70510 gke.gcr.io/prometheus-to-sd:v0.11.1-gke.1],SizeBytes:48742566,},ContainerImage{Names:[gcr.io/k8s-ingress-image-push/ingress-gce-glbc-amd64@sha256:5db27383add6d9f4ebdf0286409ac31f7f5d273690204b341a4e37998917693b gcr.io/k8s-ingress-image-push/ingress-gce-glbc-amd64:v1.20.1],SizeBytes:36598135,},ContainerImage{Names:[registry.k8s.io/addon-manager/kube-addon-manager@sha256:49cc4e6e4a3745b427ce14b0141476ab339bb65c6bc05033019e046c8727dcb0 registry.k8s.io/addon-manager/kube-addon-manager:v9.1.6],SizeBytes:30464183,},ContainerImage{Names:[registry.k8s.io/kas-network-proxy/proxy-server@sha256:2c111f004bec24888d8cfa2a812a38fb8341350abac67dcd0ac64e709dfe389c registry.k8s.io/kas-network-proxy/proxy-server:v0.0.33],SizeBytes:22020129,},ContainerImage{Names:[registry.k8s.io/metadata-proxy@sha256:e914645f22e946bce5165737e1b244e0a296ad1f0f81a9531adc57af2780978a registry.k8s.io/metadata-proxy:v0.1.12],SizeBytes:5301657,},ContainerImage{Names:[registry.k8s.io/pause@sha256:9001185023633d17a2f98ff69b6ff2615b8ea02a825adffa40422f51dfdcde9d registry.k8s.io/pause:3.8],SizeBytes:311286,},},VolumesInUse:[],VolumesAttached:[]AttachedVolume{},Config:nil,},} Nov 26 05:43:07.913: INFO: Logging kubelet events for node bootstrap-e2e-master Nov 26 05:43:07.959: INFO: Logging pods the kubelet thinks is on node bootstrap-e2e-master Nov 26 05:43:08.003: INFO: Unable to retrieve kubelet pods for node bootstrap-e2e-master: error trying to reach service: No agent available Nov 26 05:43:08.003: INFO: Logging node info for node bootstrap-e2e-minion-group-0975 Nov 26 05:43:08.045: INFO: Node Info: &Node{ObjectMeta:{bootstrap-e2e-minion-group-0975 6a0671d4-0e9d-4ac8-b359-65111262cfad 13861 0 2022-11-26 05:05:22 +0000 UTC <nil> <nil> map[beta.kubernetes.io/arch:amd64 beta.kubernetes.io/instance-type:n1-standard-2 beta.kubernetes.io/os:linux cloud.google.com/metadata-proxy-ready:true failure-domain.beta.kubernetes.io/region:us-west1 failure-domain.beta.kubernetes.io/zone:us-west1-b kubernetes.io/arch:amd64 kubernetes.io/hostname:bootstrap-e2e-minion-group-0975 kubernetes.io/os:linux node.kubernetes.io/instance-type:n1-standard-2 topology.hostpath.csi/node:bootstrap-e2e-minion-group-0975 topology.kubernetes.io/region:us-west1 topology.kubernetes.io/zone:us-west1-b] map[csi.volume.kubernetes.io/nodeid:{"csi-hostpath-multivolume-2344":"bootstrap-e2e-minion-group-0975","csi-hostpath-multivolume-5830":"bootstrap-e2e-minion-group-0975","csi-hostpath-multivolume-593":"bootstrap-e2e-minion-group-0975","csi-hostpath-provisioning-7492":"bootstrap-e2e-minion-group-0975"} node.alpha.kubernetes.io/ttl:0 volumes.kubernetes.io/controller-managed-attach-detach:true] [] [] [{kubelet Update v1 2022-11-26 05:05:22 +0000 UTC FieldsV1 {"f:metadata":{"f:annotations":{".":{},"f:volumes.kubernetes.io/controller-managed-attach-detach":{}},"f:labels":{".":{},"f:beta.kubernetes.io/arch":{},"f:beta.kubernetes.io/instance-type":{},"f:beta.kubernetes.io/os":{},"f:cloud.google.com/metadata-proxy-ready":{},"f:failure-domain.beta.kubernetes.io/region":{},"f:failure-domain.beta.kubernetes.io/zone":{},"f:kubernetes.io/arch":{},"f:kubernetes.io/hostname":{},"f:kubernetes.io/os":{},"f:node.kubernetes.io/instance-type":{},"f:topology.kubernetes.io/region":{},"f:topology.kubernetes.io/zone":{}}},"f:spec":{"f:providerID":{}}} } {kube-controller-manager Update v1 2022-11-26 05:05:23 +0000 UTC FieldsV1 {"f:metadata":{"f:annotations":{"f:node.alpha.kubernetes.io/ttl":{}}},"f:spec":{"f:podCIDR":{},"f:podCIDRs":{".":{},"v:\"10.64.2.0/24\"":{}}}} } {kube-controller-manager Update v1 2022-11-26 05:13:31 +0000 UTC FieldsV1 {"f:status":{"f:conditions":{"k:{\"type\":\"NetworkUnavailable\"}":{"f:lastHeartbeatTime":{},"f:lastTransitionTime":{},"f:message":{},"f:reason":{},"f:status":{}}}}} status} {node-problem-detector Update v1 2022-11-26 05:40:32 +0000 UTC FieldsV1 {"f:status":{"f:conditions":{"k:{\"type\":\"CorruptDockerOverlay2\"}":{".":{},"f:lastHeartbeatTime":{},"f:lastTransitionTime":{},"f:message":{},"f:reason":{},"f:status":{},"f:type":{}},"k:{\"type\":\"FrequentContainerdRestart\"}":{".":{},"f:lastHeartbeatTime":{},"f:lastTransitionTime":{},"f:message":{},"f:reason":{},"f:status":{},"f:type":{}},"k:{\"type\":\"FrequentDockerRestart\"}":{".":{},"f:lastHeartbeatTime":{},"f:lastTransitionTime":{},"f:message":{},"f:reason":{},"f:status":{},"f:type":{}},"k:{\"type\":\"FrequentKubeletRestart\"}":{".":{},"f:lastHeartbeatTime":{},"f:lastTransitionTime":{},"f:message":{},"f:reason":{},"f:status":{},"f:type":{}},"k:{\"type\":\"FrequentUnregisterNetDevice\"}":{".":{},"f:lastHeartbeatTime":{},"f:lastTransitionTime":{},"f:message":{},"f:reason":{},"f:status":{},"f:type":{}},"k:{\"type\":\"KernelDeadlock\"}":{".":{},"f:lastHeartbeatTime":{},"f:lastTransitionTime":{},"f:message":{},"f:reason":{},"f:status":{},"f:type":{}},"k:{\"type\":\"ReadonlyFilesystem\"}":{".":{},"f:lastHeartbeatTime":{},"f:lastTransitionTime":{},"f:message":{},"f:reason":{},"f:status":{},"f:type":{}}}}} status} {kubelet Update v1 2022-11-26 05:43:02 +0000 UTC FieldsV1 {"f:metadata":{"f:annotations":{"f:csi.volume.kubernetes.io/nodeid":{}},"f:labels":{"f:topology.hostpath.csi/node":{}}},"f:status":{"f:conditions":{"k:{\"type\":\"DiskPressure\"}":{"f:lastHeartbeatTime":{}},"k:{\"type\":\"MemoryPressure\"}":{"f:lastHeartbeatTime":{}},"k:{\"type\":\"PIDPressure\"}":{"f:lastHeartbeatTime":{}},"k:{\"type\":\"Ready\"}":{"f:lastHeartbeatTime":{},"f:lastTransitionTime":{},"f:message":{},"f:reason":{},"f:status":{}}},"f:images":{}}} status}]},Spec:NodeSpec{PodCIDR:10.64.2.0/24,DoNotUseExternalID:,ProviderID:gce://k8s-jkns-e2e-gce-ci-reboot/us-west1-b/bootstrap-e2e-minion-group-0975,Unschedulable:false,Taints:[]Taint{},ConfigSource:nil,PodCIDRs:[10.64.2.0/24],},Status:NodeStatus{Capacity:ResourceList{attachable-volumes-gce-pd: {{127 0} {<nil>} 127 DecimalSI},cpu: {{2 0} {<nil>} 2 DecimalSI},ephemeral-storage: {{101203873792 0} {<nil>} 98831908Ki BinarySI},hugepages-1Gi: {{0 0} {<nil>} 0 DecimalSI},hugepages-2Mi: {{0 0} {<nil>} 0 DecimalSI},memory: {{7815430144 0} {<nil>} BinarySI},pods: {{110 0} {<nil>} 110 DecimalSI},},Allocatable:ResourceList{attachable-volumes-gce-pd: {{127 0} {<nil>} 127 DecimalSI},cpu: {{2 0} {<nil>} 2 DecimalSI},ephemeral-storage: {{91083486262 0} {<nil>} 91083486262 DecimalSI},hugepages-1Gi: {{0 0} {<nil>} 0 DecimalSI},hugepages-2Mi: {{0 0} {<nil>} 0 DecimalSI},memory: {{7553286144 0} {<nil>} BinarySI},pods: {{110 0} {<nil>} 110 DecimalSI},},Phase:,Conditions:[]NodeCondition{NodeCondition{Type:CorruptDockerOverlay2,Status:False,LastHeartbeatTime:2022-11-26 05:40:32 +0000 UTC,LastTransitionTime:2022-11-26 05:05:27 +0000 UTC,Reason:NoCorruptDockerOverlay2,Message:docker overlay2 is functioning properly,},NodeCondition{Type:FrequentUnregisterNetDevice,Status:False,LastHeartbeatTime:2022-11-26 05:40:32 +0000 UTC,LastTransitionTime:2022-11-26 05:05:27 +0000 UTC,Reason:NoFrequentUnregisterNetDevice,Message:node is functioning properly,},NodeCondition{Type:FrequentKubeletRestart,Status:False,LastHeartbeatTime:2022-11-26 05:40:32 +0000 UTC,LastTransitionTime:2022-11-26 05:05:27 +0000 UTC,Reason:NoFrequentKubeletRestart,Message:kubelet is functioning properly,},NodeCondition{Type:FrequentDockerRestart,Status:False,LastHeartbeatTime:2022-11-26 05:40:32 +0000 UTC,LastTransitionTime:2022-11-26 05:05:27 +0000 UTC,Reason:NoFrequentDockerRestart,Message:docker is functioning properly,},NodeCondition{Type:FrequentContainerdRestart,Status:False,LastHeartbeatTime:2022-11-26 05:40:32 +0000 UTC,LastTransitionTime:2022-11-26 05:05:27 +0000 UTC,Reason:NoFrequentContainerdRestart,Message:containerd is functioning properly,},NodeCondition{Type:KernelDeadlock,Status:False,LastHeartbeatTime:2022-11-26 05:40:32 +0000 UTC,LastTransitionTime:2022-11-26 05:05:27 +0000 UTC,Reason:KernelHasNoDeadlock,Message:kernel has no deadlock,},NodeCondition{Type:ReadonlyFilesystem,Status:False,LastHeartbeatTime:2022-11-26 05:40:32 +0000 UTC,LastTransitionTime:2022-11-26 05:05:27 +0000 UTC,Reason:FilesystemIsNotReadOnly,Message:Filesystem is not read-only,},NodeCondition{Type:NetworkUnavailable,Status:False,LastHeartbeatTime:2022-11-26 05:05:39 +0000 UTC,LastTransitionTime:2022-11-26 05:05:39 +0000 UTC,Reason:RouteCreated,Message:RouteController created a route,},NodeCondition{Type:MemoryPressure,Status:False,LastHeartbeatTime:2022-11-26 05:39:41 +0000 UTC,LastTransitionTime:2022-11-26 05:05:22 +0000 UTC,Reason:KubeletHasSufficientMemory,Message:kubelet has sufficient memory available,},NodeCondition{Type:DiskPressure,Status:False,LastHeartbeatTime:2022-11-26 05:39:41 +0000 UTC,LastTransitionTime:2022-11-26 05:05:22 +0000 UTC,Reason:KubeletHasNoDiskPressure,Message:kubelet has no disk pressure,},NodeCondition{Type:PIDPressure,Status:False,LastHeartbeatTime:2022-11-26 05:39:41 +0000 UTC,LastTransitionTime:2022-11-26 05:05:22 +0000 UTC,Reason:KubeletHasSufficientPID,Message:kubelet has sufficient PID available,},NodeCondition{Type:Ready,Status:True,LastHeartbeatTime:2022-11-26 05:39:41 +0000 UTC,LastTransitionTime:2022-11-26 05:05:23 +0000 UTC,Reason:KubeletReady,Message:kubelet is posting ready status. AppArmor enabled,},},Addresses:[]NodeAddress{NodeAddress{Type:InternalIP,Address:10.138.0.5,},NodeAddress{Type:ExternalIP,Address:35.247.104.236,},NodeAddress{Type:InternalDNS,Address:bootstrap-e2e-minion-group-0975.c.k8s-jkns-e2e-gce-ci-reboot.internal,},NodeAddress{Type:Hostname,Address:bootstrap-e2e-minion-group-0975.c.k8s-jkns-e2e-gce-ci-reboot.internal,},},DaemonEndpoints:NodeDaemonEndpoints{KubeletEndpoint:DaemonEndpoint{Port:10250,},},NodeInfo:NodeSystemInfo{MachineID:b38385f7976a474936da82f5f565760a,SystemUUID:b38385f7-976a-4749-36da-82f5f565760a,BootID:0aa74273-9b8d-48ac-bad0-0a1805f010ea,KernelVersion:5.10.123+,OSImage:Container-Optimized OS from Google,ContainerRuntimeVersion:containerd://1.7.0-beta.0-149-gd06318622,KubeletVersion:v1.27.0-alpha.0.50+70617042976dc1,KubeProxyVersion:v1.27.0-alpha.0.50+70617042976dc1,OperatingSystem:linux,Architecture:amd64,},Images:[]ContainerImage{ContainerImage{Names:[registry.k8s.io/e2e-test-images/jessie-dnsutils@sha256:24aaf2626d6b27864c29de2097e8bbb840b3a414271bf7c8995e431e47d8408e registry.k8s.io/e2e-test-images/jessie-dnsutils:1.7],SizeBytes:112030336,},ContainerImage{Names:[registry.k8s.io/e2e-test-images/volume/nfs@sha256:3bda73f2428522b0e342af80a0b9679e8594c2126f2b3cca39ed787589741b9e registry.k8s.io/e2e-test-images/volume/nfs:1.3],SizeBytes:95836203,},ContainerImage{Names:[registry.k8s.io/sig-storage/nfs-provisioner@sha256:e943bb77c7df05ebdc8c7888b2db289b13bf9f012d6a3a5a74f14d4d5743d439 registry.k8s.io/sig-storage/nfs-provisioner:v3.0.1],SizeBytes:90632047,},ContainerImage{Names:[registry.k8s.io/kube-proxy-amd64:v1.27.0-alpha.0.50_70617042976dc1],SizeBytes:67201736,},ContainerImage{Names:[registry.k8s.io/e2e-test-images/agnhost@sha256:16bbf38c463a4223d8cfe4da12bc61010b082a79b4bb003e2d3ba3ece5dd5f9e registry.k8s.io/e2e-test-images/agnhost:2.43],SizeBytes:51706353,},ContainerImage{Names:[gke.gcr.io/prometheus-to-sd@sha256:e739643c3939ba0b161425f45a1989eedfc4a3b166db9a7100863296b4c70510 gke.gcr.io/prometheus-to-sd:v0.11.1-gke.1],SizeBytes:48742566,},ContainerImage{Names:[registry.k8s.io/e2e-test-images/httpd@sha256:148b022f5c5da426fc2f3c14b5c0867e58ef05961510c84749ac1fddcb0fef22 registry.k8s.io/e2e-test-images/httpd:2.4.38-4],SizeBytes:40764257,},ContainerImage{Names:[registry.k8s.io/metrics-server/metrics-server@sha256:6385aec64bb97040a5e692947107b81e178555c7a5b71caa90d733e4130efc10 registry.k8s.io/metrics-server/metrics-server:v0.5.2],SizeBytes:26023008,},ContainerImage{Names:[registry.k8s.io/sig-storage/csi-provisioner@sha256:ee3b525d5b89db99da3b8eb521d9cd90cb6e9ef0fbb651e98bb37be78d36b5b8 registry.k8s.io/sig-storage/csi-provisioner:v3.3.0],SizeBytes:25491225,},ContainerImage{Names:[registry.k8s.io/sig-storage/csi-resizer@sha256:425d8f1b769398127767b06ed97ce62578a3179bcb99809ce93a1649e025ffe7 registry.k8s.io/sig-storage/csi-resizer:v1.6.0],SizeBytes:24148884,},ContainerImage{Names:[registry.k8s.io/sig-storage/csi-snapshotter@sha256:291334908ddf71a4661fd7f6d9d97274de8a5378a2b6fdfeb2ce73414a34f82f registry.k8s.io/sig-storage/csi-snapshotter:v6.1.0],SizeBytes:23881995,},ContainerImage{Names:[registry.k8s.io/sig-storage/csi-attacher@sha256:9a685020911e2725ad019dbce6e4a5ab93d51e3d4557f115e64343345e05781b registry.k8s.io/sig-storage/csi-attacher:v4.0.0],SizeBytes:23847201,},ContainerImage{Names:[registry.k8s.io/sig-storage/hostpathplugin@sha256:92257881c1d6493cf18299a24af42330f891166560047902b8d431fb66b01af5 registry.k8s.io/sig-storage/hostpathplugin:v1.9.0],SizeBytes:18758628,},ContainerImage{Names:[registry.k8s.io/autoscaling/addon-resizer@sha256:43f129b81d28f0fdd54de6d8e7eacd5728030782e03db16087fc241ad747d3d6 registry.k8s.io/autoscaling/addon-resizer:1.8.14],SizeBytes:10153852,},ContainerImage{Names:[registry.k8s.io/sig-storage/csi-node-driver-registrar@sha256:0103eee7c35e3e0b5cd8cdca9850dc71c793cdeb6669d8be7a89440da2d06ae4 registry.k8s.io/sig-storage/csi-node-driver-registrar:v2.5.1],SizeBytes:9133109,},ContainerImage{Names:[registry.k8s.io/sig-storage/livenessprobe@sha256:933940f13b3ea0abc62e656c1aa5c5b47c04b15d71250413a6b821bd0c58b94e registry.k8s.io/sig-storage/livenessprobe:v2.7.0],SizeBytes:8688564,},ContainerImage{Names:[registry.k8s.io/kas-network-proxy/proxy-agent@sha256:48f2a4ec3e10553a81b8dd1c6fa5fe4bcc9617f78e71c1ca89c6921335e2d7da registry.k8s.io/kas-network-proxy/proxy-agent:v0.0.33],SizeBytes:8512162,},ContainerImage{Names:[registry.k8s.io/metadata-proxy@sha256:e914645f22e946bce5165737e1b244e0a296ad1f0f81a9531adc57af2780978a registry.k8s.io/metadata-proxy:v0.1.12],SizeBytes:5301657,},ContainerImage{Names:[registry.k8s.io/e2e-test-images/busybox@sha256:c318242786b139d18676b1c09a0ad7f15fc17f8f16a5b2e625cd0dc8c9703daf registry.k8s.io/e2e-test-images/busybox:1.29-2],SizeBytes:732424,},ContainerImage{Names:[registry.k8s.io/e2e-test-images/busybox@sha256:2e0f836850e09b8b7cc937681d6194537a09fbd5f6b9e08f4d646a85128e8937 registry.k8s.io/e2e-test-images/busybox:1.29-4],SizeBytes:731990,},ContainerImage{Names:[registry.k8s.io/pause@sha256:9001185023633d17a2f98ff69b6ff2615b8ea02a825adffa40422f51dfdcde9d registry.k8s.io/pause:3.8],SizeBytes:311286,},},VolumesInUse:[],VolumesAttached:[]AttachedVolume{},Config:nil,},} Nov 26 05:43:08.045: INFO: Logging kubelet events for node bootstrap-e2e-minion-group-0975 Nov 26 05:43:08.090: INFO: Logging pods the kubelet thinks is on node bootstrap-e2e-minion-group-0975 Nov 26 05:43:08.148: INFO: Unable to retrieve kubelet pods for node bootstrap-e2e-minion-group-0975: error trying to reach service: No agent available Nov 26 05:43:08.148: INFO: Logging node info for node bootstrap-e2e-minion-group-rwsn Nov 26 05:43:08.190: INFO: Node Info: &Node{ObjectMeta:{bootstrap-e2e-minion-group-rwsn adedf225-b462-4305-a60c-9042d31ca1d2 13880 0 2022-11-26 05:05:15 +0000 UTC <nil> <nil> map[beta.kubernetes.io/arch:amd64 beta.kubernetes.io/instance-type:n1-standard-2 beta.kubernetes.io/os:linux cloud.google.com/metadata-proxy-ready:true failure-domain.beta.kubernetes.io/region:us-west1 failure-domain.beta.kubernetes.io/zone:us-west1-b kubernetes.io/arch:amd64 kubernetes.io/hostname:bootstrap-e2e-minion-group-rwsn kubernetes.io/os:linux node.kubernetes.io/instance-type:n1-standard-2 topology.hostpath.csi/node:bootstrap-e2e-minion-group-rwsn topology.kubernetes.io/region:us-west1 topology.kubernetes.io/zone:us-west1-b] map[node.alpha.kubernetes.io/ttl:0 volumes.kubernetes.io/controller-managed-attach-detach:true] [] [] [{kubelet Update v1 2022-11-26 05:05:15 +0000 UTC FieldsV1 {"f:metadata":{"f:annotations":{".":{},"f:volumes.kubernetes.io/controller-managed-attach-detach":{}},"f:labels":{".":{},"f:beta.kubernetes.io/arch":{},"f:beta.kubernetes.io/instance-type":{},"f:beta.kubernetes.io/os":{},"f:cloud.google.com/metadata-proxy-ready":{},"f:failure-domain.beta.kubernetes.io/region":{},"f:failure-domain.beta.kubernetes.io/zone":{},"f:kubernetes.io/arch":{},"f:kubernetes.io/hostname":{},"f:kubernetes.io/os":{},"f:node.kubernetes.io/instance-type":{},"f:topology.kubernetes.io/region":{},"f:topology.kubernetes.io/zone":{}}},"f:spec":{"f:providerID":{}}} } {kube-controller-manager Update v1 2022-11-26 05:05:1