This job view page is being replaced by Spyglass soon. Check out the new job view.
ResultFAILURE
Tests 205 failed / 137 succeeded
Started2022-11-26 20:35
Elapsed1h13m
Revision
Buildercc10ec08-6dc9-11ed-a6da-22215f11743a
control_plane_node_os_imagecos-97-16919-103-16
infra-commitb02ce7741
job-versionv1.27.0-alpha.0.50+70617042976dc1
kubetest-versionv20221116-7c85504268
repogithub.com/containerd/containerd
repo-commitd06318622178a2209338d7f57eb8800b9f2e72fb
repos{u'github.com/containerd/containerd': u'main'}
revisionv1.27.0-alpha.0.50+70617042976dc1
worker_node_os_imagecos-97-16919-103-16

Test Failures


Kubernetes e2e suite [It] [sig-api-machinery] Servers with support for API chunking should support continue listing from the last key if the original version has been compacted away, though the list is inconsistent [Slow] 3m58s

go run hack/e2e.go -v --test --test_args='--ginkgo.focus=Kubernetes\se2e\ssuite\s\[It\]\s\[sig\-api\-machinery\]\sServers\swith\ssupport\sfor\sAPI\schunking\sshould\ssupport\scontinue\slisting\sfrom\sthe\slast\skey\sif\sthe\soriginal\sversion\shas\sbeen\scompacted\saway\,\sthough\sthe\slist\sis\sinconsistent\s\[Slow\]$'
test/e2e/apimachinery/chunking.go:177
k8s.io/kubernetes/test/e2e/apimachinery.glob..func4.3()
	test/e2e/apimachinery/chunking.go:177 +0x7fc
There were additional failures detected after the initial failure:
[FAILED]
Nov 26 21:27:58.334: failed to list events in namespace "chunking-2842": Get "https://35.233.174.213/api/v1/namespaces/chunking-2842/events": dial tcp 35.233.174.213:443: connect: connection refused
In [DeferCleanup (Each)] at: test/e2e/framework/debug/dump.go:44
----------
[FAILED]
Nov 26 21:27:58.374: Couldn't delete ns: "chunking-2842": Delete "https://35.233.174.213/api/v1/namespaces/chunking-2842": dial tcp 35.233.174.213:443: connect: connection refused (&url.Error{Op:"Delete", URL:"https://35.233.174.213/api/v1/namespaces/chunking-2842", Err:(*net.OpError)(0xc003cc6960)})
In [DeferCleanup (Each)] at: test/e2e/framework/framework.go:370

				
				Click to see stdout/stderrfrom junit_01.xml

Filter through log files | View test history on testgrid


Kubernetes e2e suite [It] [sig-apps] StatefulSet Basic StatefulSet functionality [StatefulSetBasic] Burst scaling should run to completion even with unhealthy pods [Slow] [Conformance] 14m5s

go run hack/e2e.go -v --test --test_args='--ginkgo.focus=Kubernetes\se2e\ssuite\s\[It\]\s\[sig\-apps\]\sStatefulSet\sBasic\sStatefulSet\sfunctionality\s\[StatefulSetBasic\]\sBurst\sscaling\sshould\srun\sto\scompletion\seven\swith\sunhealthy\spods\s\[Slow\]\s\[Conformance\]$'
test/e2e/framework/statefulset/rest.go:69
k8s.io/kubernetes/test/e2e/framework/statefulset.GetPodList({0x801de88, 0xc003f24b60}, 0xc001745900)
	test/e2e/framework/statefulset/rest.go:69 +0x153
k8s.io/kubernetes/test/e2e/framework/statefulset.WaitForRunning.func1()
	test/e2e/framework/statefulset/wait.go:37 +0x4a
k8s.io/kubernetes/vendor/k8s.io/apimachinery/pkg/util/wait.ConditionFunc.WithContext.func1({0x2742871, 0x0})
	vendor/k8s.io/apimachinery/pkg/util/wait/wait.go:222 +0x1b
k8s.io/kubernetes/vendor/k8s.io/apimachinery/pkg/util/wait.runConditionWithCrashProtectionWithContext({0x7fe0bc8?, 0xc0000820c8?}, 0xc0001d0380?)
	vendor/k8s.io/apimachinery/pkg/util/wait/wait.go:235 +0x57
k8s.io/kubernetes/vendor/k8s.io/apimachinery/pkg/util/wait.WaitForWithContext({0x7fe0bc8, 0xc0000820c8}, 0xc001f74570, 0x2fdb16a?)
	vendor/k8s.io/apimachinery/pkg/util/wait/wait.go:662 +0x10c
k8s.io/kubernetes/vendor/k8s.io/apimachinery/pkg/util/wait.poll({0x7fe0bc8, 0xc0000820c8}, 0xf8?, 0x2fd9d05?, 0x20?)
	vendor/k8s.io/apimachinery/pkg/util/wait/wait.go:596 +0x9a
k8s.io/kubernetes/vendor/k8s.io/apimachinery/pkg/util/wait.PollImmediateWithContext({0x7fe0bc8, 0xc0000820c8}, 0x1?, 0xc002e7fe48?, 0x262a967?)
	vendor/k8s.io/apimachinery/pkg/util/wait/wait.go:528 +0x4a
k8s.io/kubernetes/vendor/k8s.io/apimachinery/pkg/util/wait.PollImmediate(0x801de88?, 0xc003f24b60?, 0xc002e7fe88?)
	vendor/k8s.io/apimachinery/pkg/util/wait/wait.go:514 +0x50
k8s.io/kubernetes/test/e2e/framework/statefulset.WaitForRunning({0x801de88?, 0xc003f24b60}, 0x3, 0x3, 0xc001745900)
	test/e2e/framework/statefulset/wait.go:35 +0xbd
k8s.io/kubernetes/test/e2e/framework/statefulset.WaitForRunningAndReady(...)
	test/e2e/framework/statefulset/wait.go:80
k8s.io/kubernetes/test/e2e/apps.glob..func10.2.11()
	test/e2e/apps/statefulset.go:719 +0x3d0
There were additional failures detected after the initial failure:
[FAILED]
Nov 26 20:57:15.373: Get "https://35.233.174.213/apis/apps/v1/namespaces/statefulset-4883/statefulsets": stream error: stream ID 33; INTERNAL_ERROR; received from peer
In [AfterEach] at: test/e2e/framework/statefulset/rest.go:76

				
				Click to see stdout/stderrfrom junit_01.xml

Filter through log files | View test history on testgrid


Kubernetes e2e suite [It] [sig-apps] StatefulSet Basic StatefulSet functionality [StatefulSetBasic] Scaling should happen in predictable order and halt if any stateful pod is unhealthy [Slow] [Conformance] 8m41s

go run hack/e2e.go -v --test --test_args='--ginkgo.focus=Kubernetes\se2e\ssuite\s\[It\]\s\[sig\-apps\]\sStatefulSet\sBasic\sStatefulSet\sfunctionality\s\[StatefulSetBasic\]\sScaling\sshould\shappen\sin\spredictable\sorder\sand\shalt\sif\sany\sstateful\spod\sis\sunhealthy\s\[Slow\]\s\[Conformance\]$'
test/e2e/framework/statefulset/rest.go:69
k8s.io/kubernetes/test/e2e/framework/statefulset.GetPodList({0x801de88, 0xc003c7a000}, 0xc00102cf00)
	test/e2e/framework/statefulset/rest.go:69 +0x153
k8s.io/kubernetes/test/e2e/framework/statefulset.WaitForRunning.func1()
	test/e2e/framework/statefulset/wait.go:37 +0x4a
k8s.io/kubernetes/vendor/k8s.io/apimachinery/pkg/util/wait.ConditionFunc.WithContext.func1({0x2742871, 0x0})
	vendor/k8s.io/apimachinery/pkg/util/wait/wait.go:222 +0x1b
k8s.io/kubernetes/vendor/k8s.io/apimachinery/pkg/util/wait.runConditionWithCrashProtectionWithContext({0x7fe0bc8?, 0xc000136000?}, 0xc003156a20?)
	vendor/k8s.io/apimachinery/pkg/util/wait/wait.go:235 +0x57
k8s.io/kubernetes/vendor/k8s.io/apimachinery/pkg/util/wait.WaitForWithContext({0x7fe0bc8, 0xc000136000}, 0xc00016db60, 0x2fdb16a?)
	vendor/k8s.io/apimachinery/pkg/util/wait/wait.go:662 +0x10c
k8s.io/kubernetes/vendor/k8s.io/apimachinery/pkg/util/wait.poll({0x7fe0bc8, 0xc000136000}, 0x90?, 0x2fd9d05?, 0x20?)
	vendor/k8s.io/apimachinery/pkg/util/wait/wait.go:596 +0x9a
k8s.io/kubernetes/vendor/k8s.io/apimachinery/pkg/util/wait.PollImmediateWithContext({0x7fe0bc8, 0xc000136000}, 0x65cbc00?, 0xc0036fbde0?, 0x262a967?)
	vendor/k8s.io/apimachinery/pkg/util/wait/wait.go:528 +0x4a
k8s.io/kubernetes/vendor/k8s.io/apimachinery/pkg/util/wait.PollImmediate(0x277?, 0x0?, 0x0?)
	vendor/k8s.io/apimachinery/pkg/util/wait/wait.go:514 +0x50
k8s.io/kubernetes/test/e2e/framework/statefulset.WaitForRunning({0x801de88?, 0xc003c7a000}, 0x1, 0x1, 0xc00102cf00)
	test/e2e/framework/statefulset/wait.go:35 +0xbd
k8s.io/kubernetes/test/e2e/framework/statefulset.WaitForRunningAndReady(...)
	test/e2e/framework/statefulset/wait.go:80
k8s.io/kubernetes/test/e2e/apps.glob..func10.2.10()
	test/e2e/apps/statefulset.go:632 +0x57b
				
				Click to see stdout/stderrfrom junit_01.xml

Filter through log files | View test history on testgrid


Kubernetes e2e suite [It] [sig-auth] ServiceAccounts should support InClusterConfig with token rotation [Slow] 20m8s

go run hack/e2e.go -v --test --test_args='--ginkgo.focus=Kubernetes\se2e\ssuite\s\[It\]\s\[sig\-auth\]\sServiceAccounts\sshould\ssupport\sInClusterConfig\swith\stoken\srotation\s\[Slow\]$'
test/e2e/auth/service_accounts.go:520
k8s.io/kubernetes/test/e2e/auth.glob..func5.6()
	test/e2e/auth/service_accounts.go:520 +0x9ab
				
				Click to see stdout/stderrfrom junit_01.xml

Filter through log files | View test history on testgrid


Kubernetes e2e suite [It] [sig-cli] Kubectl client Simple pod should return command exit codes [Slow] running a failing command without --restart=Never, but with --rm 6m32s

go run hack/e2e.go -v --test --test_args='--ginkgo.focus=Kubernetes\se2e\ssuite\s\[It\]\s\[sig\-cli\]\sKubectl\sclient\sSimple\spod\sshould\sreturn\scommand\sexit\scodes\s\[Slow\]\srunning\sa\sfailing\scommand\swithout\s\-\-restart\=Never\,\sbut\swith\s\-\-rm$'
test/e2e/kubectl/kubectl.go:415
k8s.io/kubernetes/test/e2e/kubectl.glob..func1.8.1()
	test/e2e/kubectl/kubectl.go:415 +0x245
There were additional failures detected after the initial failure:
[FAILED]
Nov 26 21:06:13.967: error running /workspace/github.com/containerd/containerd/kubernetes/platforms/linux/amd64/kubectl --server=https://35.233.174.213 --kubeconfig=/workspace/.kube/config --namespace=kubectl-4639 delete --grace-period=0 --force -f -:
Command stdout:

stderr:
Warning: Immediate deletion does not wait for confirmation that the running resource has been terminated. The resource may continue to run on the cluster indefinitely.
Error from server (Timeout): error when deleting "STDIN": Timeout: request did not complete within requested timeout - context deadline exceeded

error:
exit status 1
In [AfterEach] at: test/e2e/framework/kubectl/builder.go:87

				
				Click to see stdout/stderrfrom junit_01.xml

Filter through log files | View test history on testgrid


Kubernetes e2e suite [It] [sig-cloud-provider-gcp] Addon update should propagate add-on file changes [Slow] 6m43s

go run hack/e2e.go -v --test --test_args='--ginkgo.focus=Kubernetes\se2e\ssuite\s\[It\]\s\[sig\-cloud\-provider\-gcp\]\sAddon\supdate\sshould\spropagate\sadd\-on\sfile\schanges\s\[Slow\]$'
test/e2e/cloud/gcp/addon_update.go:353
k8s.io/kubernetes/test/e2e/cloud/gcp.waitForReplicationControllerInAddonTest({0x801de88?, 0xc004fd29c0?}, {0x75ce977?, 0x4?}, {0x760025e?, 0xc003ef5e30?}, 0x1d?)
	test/e2e/cloud/gcp/addon_update.go:353 +0x54
k8s.io/kubernetes/test/e2e/cloud/gcp.glob..func1.3()
	test/e2e/cloud/gcp/addon_update.go:311 +0x1025
				
				Click to see stdout/stderrfrom junit_01.xml

Filter through log files | View test history on testgrid


Kubernetes e2e suite [It] [sig-network] LoadBalancers ESIPP [Slow] should handle updates to ExternalTrafficPolicy field 7m23s

go run hack/e2e.go -v --test --test_args='--ginkgo.focus=Kubernetes\se2e\ssuite\s\[It\]\s\[sig\-network\]\sLoadBalancers\sESIPP\s\[Slow\]\sshould\shandle\supdates\sto\sExternalTrafficPolicy\sfield$'
test/e2e/framework/network/utils.go:866
k8s.io/kubernetes/test/e2e/framework/network.(*NetworkingTestConfig).createNetProxyPods(0xc0002a6620, {0x75c6f7c, 0x9}, 0xc00271b3b0)
	test/e2e/framework/network/utils.go:866 +0x1d0
k8s.io/kubernetes/test/e2e/framework/network.(*NetworkingTestConfig).setupCore(0xc0002a6620, 0x7fbda0b04a30?)
	test/e2e/framework/network/utils.go:763 +0x55
k8s.io/kubernetes/test/e2e/framework/network.(*NetworkingTestConfig).setup(0xc0002a6620, 0x3c?)
	test/e2e/framework/network/utils.go:778 +0x3e
k8s.io/kubernetes/test/e2e/framework/network.NewNetworkingTestConfig(0xc001110000, {0x0, 0x0, 0x7f8f6d0?})
	test/e2e/framework/network/utils.go:131 +0x125
k8s.io/kubernetes/test/e2e/network.glob..func20.7()
	test/e2e/network/loadbalancer.go:1544 +0x417
				
				Click to see stdout/stderrfrom junit_01.xml

Filter through log files | View test history on testgrid


Kubernetes e2e suite [It] [sig-network] LoadBalancers ESIPP [Slow] should only target nodes with endpoints 3m46s

go run hack/e2e.go -v --test --test_args='--ginkgo.focus=Kubernetes\se2e\ssuite\s\[It\]\s\[sig\-network\]\sLoadBalancers\sESIPP\s\[Slow\]\sshould\sonly\starget\snodes\swith\sendpoints$'
test/e2e/framework/network/utils.go:866
k8s.io/kubernetes/test/e2e/framework/network.(*NetworkingTestConfig).createNetProxyPods(0xc000330d20, {0x75c6f7c, 0x9}, 0xc00383f9b0)
	test/e2e/framework/network/utils.go:866 +0x1d0
k8s.io/kubernetes/test/e2e/framework/network.(*NetworkingTestConfig).setupCore(0xc000330d20, 0x7f948ced3448?)
	test/e2e/framework/network/utils.go:763 +0x55
k8s.io/kubernetes/test/e2e/framework/network.(*NetworkingTestConfig).setup(0xc000330d20, 0x3c?)
	test/e2e/framework/network/utils.go:778 +0x3e
k8s.io/kubernetes/test/e2e/framework/network.NewNetworkingTestConfig(0xc0012fa000, {0x0, 0x0, 0xc00011bd20?})
	test/e2e/framework/network/utils.go:131 +0x125
k8s.io/kubernetes/test/e2e/network.glob..func20.5()
	test/e2e/network/loadbalancer.go:1382 +0x445
There were additional failures detected after the initial failure:
[FAILED]
Nov 26 21:27:48.344: failed to list events in namespace "esipp-1895": Get "https://35.233.174.213/api/v1/namespaces/esipp-1895/events": dial tcp 35.233.174.213:443: connect: connection refused
In [DeferCleanup (Each)] at: test/e2e/framework/debug/dump.go:44
----------
[FAILED]
Nov 26 21:27:48.384: Couldn't delete ns: "esipp-1895": Delete "https://35.233.174.213/api/v1/namespaces/esipp-1895": dial tcp 35.233.174.213:443: connect: connection refused (&url.Error{Op:"Delete", URL:"https://35.233.174.213/api/v1/namespaces/esipp-1895", Err:(*net.OpError)(0xc002d6d9f0)})
In [DeferCleanup (Each)] at: test/e2e/framework/framework.go:370

				
				Click to see stdout/stderrfrom junit_01.xml

Filter through log files | View test history on testgrid


Kubernetes e2e suite [It] [sig-network] LoadBalancers ESIPP [Slow] should work for type=LoadBalancer 8m25s

go run hack/e2e.go -v --test --test_args='--ginkgo.focus=Kubernetes\se2e\ssuite\s\[It\]\s\[sig\-network\]\sLoadBalancers\sESIPP\s\[Slow\]\sshould\swork\sfor\stype\=LoadBalancer$'
test/e2e/framework/network/utils.go:866
k8s.io/kubernetes/test/e2e/framework/network.(*NetworkingTestConfig).createNetProxyPods(0xc000c56380, {0x75c6f7c, 0x9}, 0xc001b99ce0)
	test/e2e/framework/network/utils.go:866 +0x1d0
k8s.io/kubernetes/test/e2e/framework/network.(*NetworkingTestConfig).setupCore(0xc000c56380, 0x7fbda13c4fd0?)
	test/e2e/framework/network/utils.go:763 +0x55
k8s.io/kubernetes/test/e2e/framework/network.(*NetworkingTestConfig).setup(0xc000c56380, 0x3c?)
	test/e2e/framework/network/utils.go:778 +0x3e
k8s.io/kubernetes/test/e2e/framework/network.NewNetworkingTestConfig(0xc001110000, {0x0, 0x0, 0x0?})
	test/e2e/framework/network/utils.go:131 +0x125
k8s.io/kubernetes/test/e2e/network.glob..func20.3.1()
	test/e2e/network/loadbalancer.go:1285 +0x10a
k8s.io/kubernetes/test/e2e/network.glob..func20.3()
	test/e2e/network/loadbalancer.go:1312 +0x37f
				
				Click to see stdout/stderrfrom junit_01.xml

Filter through log files | View test history on testgrid


Kubernetes e2e suite [It] [sig-network] LoadBalancers ESIPP [Slow] should work for type=NodePort 5m11s

go run hack/e2e.go -v --test --test_args='--ginkgo.focus=Kubernetes\se2e\ssuite\s\[It\]\s\[sig\-network\]\sLoadBalancers\sESIPP\s\[Slow\]\sshould\swork\sfor\stype\=NodePort$'
test/e2e/framework/pod/exec_util.go:126
k8s.io/kubernetes/test/e2e/framework/pod.execCommandInPodWithFullOutput(0x7775853?, {0xc001fac1c8, 0x12}, {0xc0034e6cf0, 0x3, 0x3})
	test/e2e/framework/pod/exec_util.go:126 +0x133
k8s.io/kubernetes/test/e2e/framework/pod.ExecShellInPodWithFullOutput(...)
	test/e2e/framework/pod/exec_util.go:138
k8s.io/kubernetes/test/e2e/framework/network.(*NetworkingTestConfig).GetResponseFromContainer(0xc00142e9a0, {0x75b767e, 0x4}, {0x75c2d47, 0x8}, {0xc0012d6490, 0xb}, {0xc004ca94c0, 0xa}, 0x2378, ...)
	test/e2e/framework/network/utils.go:396 +0x32a
k8s.io/kubernetes/test/e2e/framework/network.(*NetworkingTestConfig).GetResponseFromTestContainer(...)
	test/e2e/framework/network/utils.go:411
k8s.io/kubernetes/test/e2e/network.GetHTTPContentFromTestContainer.func1()
	test/e2e/network/util.go:62 +0x91
k8s.io/kubernetes/vendor/k8s.io/apimachinery/pkg/util/wait.ConditionFunc.WithContext.func1({0x2742871, 0x0})
	vendor/k8s.io/apimachinery/pkg/util/wait/wait.go:222 +0x1b
k8s.io/kubernetes/vendor/k8s.io/apimachinery/pkg/util/wait.runConditionWithCrashProtectionWithContext({0x7fe0bc8?, 0xc0001b0000?}, 0x0?)
	vendor/k8s.io/apimachinery/pkg/util/wait/wait.go:235 +0x57
k8s.io/kubernetes/vendor/k8s.io/apimachinery/pkg/util/wait.WaitForWithContext({0x7fe0bc8, 0xc0001b0000}, 0xc000fcb038, 0x2fdb16a?)
	vendor/k8s.io/apimachinery/pkg/util/wait/wait.go:662 +0x10c
k8s.io/kubernetes/vendor/k8s.io/apimachinery/pkg/util/wait.poll({0x7fe0bc8, 0xc0001b0000}, 0x58?, 0x2fd9d05?, 0x40?)
	vendor/k8s.io/apimachinery/pkg/util/wait/wait.go:596 +0x9a
k8s.io/kubernetes/vendor/k8s.io/apimachinery/pkg/util/wait.PollImmediateWithContext({0x7fe0bc8, 0xc0001b0000}, 0xc001671aa0?, 0xc001c63da8?, 0x262a967?)
	vendor/k8s.io/apimachinery/pkg/util/wait/wait.go:528 +0x4a
k8s.io/kubernetes/vendor/k8s.io/apimachinery/pkg/util/wait.PollImmediate(0x0?, 0x0?, 0xc0d8bed39d412f53?)
	vendor/k8s.io/apimachinery/pkg/util/wait/wait.go:514 +0x50
k8s.io/kubernetes/test/e2e/network.GetHTTPContentFromTestContainer(0xc00142e9a0, {0xc004ca94c0, 0xa}, 0x7a49, 0x4?, {0x75c2d47, 0x8})
	test/e2e/network/util.go:69 +0x125
k8s.io/kubernetes/test/e2e/network.glob..func20.4()
	test/e2e/network/loadbalancer.go:1336 +0x2dc
There were additional failures detected after the initial failure:
[FAILED]
Nov 26 21:27:48.978: failed to list events in namespace "esipp-6606": Get "https://35.233.174.213/api/v1/namespaces/esipp-6606/events": dial tcp 35.233.174.213:443: connect: connection refused
In [DeferCleanup (Each)] at: test/e2e/framework/debug/dump.go:44
----------
[FAILED]
Nov 26 21:27:49.018: Couldn't delete ns: "esipp-6606": Delete "https://35.233.174.213/api/v1/namespaces/esipp-6606": dial tcp 35.233.174.213:443: connect: connection refused (&url.Error{Op:"Delete", URL:"https://35.233.174.213/api/v1/namespaces/esipp-6606", Err:(*net.OpError)(0xc0031b2000)})
In [DeferCleanup (Each)] at: test/e2e/framework/framework.go:370

				
				Click to see stdout/stderrfrom junit_01.xml

Filter through log files | View test history on testgrid


Kubernetes e2e suite [It] [sig-network] LoadBalancers ESIPP [Slow] should work from pods 10m36s

go run hack/e2e.go -v --test --test_args='--ginkgo.focus=Kubernetes\se2e\ssuite\s\[It\]\s\[sig\-network\]\sLoadBalancers\sESIPP\s\[Slow\]\sshould\swork\sfrom\spods$'
test/e2e/network/loadbalancer.go:1429
k8s.io/kubernetes/test/e2e/network.glob..func20.6()
	test/e2e/network/loadbalancer.go:1429 +0xdd