This job view page is being replaced by Spyglass soon. Check out the new job view.
ResultFAILURE
Tests 59 failed / 616 succeeded
Started2019-11-11 04:07
Elapsed2h35m
Revision
Buildergke-prow-ssd-pool-1a225945-jhsp
links{u'resultstore': {u'url': u'https://source.cloud.google.com/results/invocations/0d7efec0-5aa4-484b-96b2-cbe431a03932/targets/test'}}
podbe6a9ddb-0438-11ea-88c9-c26ecbf388b3
resultstorehttps://source.cloud.google.com/results/invocations/0d7efec0-5aa4-484b-96b2-cbe431a03932/targets/test
infra-commit830666073
job-versionv1.16.3-beta.0.56+b3cbbae08ec52a-dirty
podbe6a9ddb-0438-11ea-88c9-c26ecbf388b3
repok8s.io/kubernetes
repo-commitb3cbbae08ec52a7fc73d334838e18d17e8512749
repos{u'k8s.io/kubernetes': u'release-1.16', u'sigs.k8s.io/cloud-provider-azure': u'master'}
revisionv1.16.3-beta.0.56+b3cbbae08ec52a-dirty

Test Failures


Kubernetes e2e suite [k8s.io] Container Lifecycle Hook when create a pod with lifecycle hook should execute prestop http hook properly [NodeConformance] [Conformance] 14m51s

go run hack/e2e.go -v --test --test_args='--ginkgo.focus=Kubernetes\se2e\ssuite\s\[k8s\.io\]\sContainer\sLifecycle\sHook\swhen\screate\sa\spod\swith\slifecycle\shook\sshould\sexecute\sprestop\shttp\shook\sproperly\s\[NodeConformance\]\s\[Conformance\]$'
test/e2e/framework/framework.go:698
Nov 11 05:07:11.651: wait for pod "pod-with-prestop-http-hook" to disappear
Expected success, but got an error:
    <*errors.errorString | 0xc0000d5090>: {
        s: "timed out waiting for the condition",
    }
    timed out waiting for the condition
test/e2e/framework/pods.go:178
				
				Click to see stdout/stderrfrom junit_07.xml

Find pod-with-prestop-http-hook mentions in log files | View test history on testgrid


Kubernetes e2e suite [k8s.io] Pods should support pod readiness gates [NodeFeature:PodReadinessGate] 1m25s

go run hack/e2e.go -v --test --test_args='--ginkgo.focus=Kubernetes\se2e\ssuite\s\[k8s\.io\]\sPods\sshould\ssupport\spod\sreadiness\sgates\s\[NodeFeature\:PodReadinessGate\]$'
test/e2e/common/pods.go:777
Nov 11 04:44:54.818: Unexpected error:
    <*errors.errorString | 0xc0000d5090>: {
        s: "timed out waiting for the condition",
    }
    timed out waiting for the condition
occurred
test/e2e/common/pods.go:811
				
				Click to see stdout/stderrfrom junit_03.xml

Filter through log files | View test history on testgrid


Kubernetes e2e suite [k8s.io] Pods should support pod readiness gates [NodeFeature:PodReadinessGate] 1m40s

go run hack/e2e.go -v --test --test_args='--ginkgo.focus=Kubernetes\se2e\ssuite\s\[k8s\.io\]\sPods\sshould\ssupport\spod\sreadiness\sgates\s\[NodeFeature\:PodReadinessGate\]$'
test/e2e/common/pods.go:777
Nov 11 04:46:30.043: Unexpected error:
    <*errors.errorString | 0xc0000d5090>: {
        s: "timed out waiting for the condition",
    }
    timed out waiting for the condition
occurred
test/e2e/common/pods.go:811
				
				Click to see stdout/stderrfrom junit_03.xml

Filter through log files | View test history on testgrid


Kubernetes e2e suite [k8s.io] [sig-node] Events should be sent by kubelets and the scheduler about pods scheduling and running [Conformance] 14m29s

go run hack/e2e.go -v --test --test_args='--ginkgo.focus=Kubernetes\se2e\ssuite\s\[k8s\.io\]\s\[sig\-node\]\sEvents\sshould\sbe\ssent\sby\skubelets\sand\sthe\sscheduler\sabout\spods\sscheduling\sand\srunning\s\s\[Conformance\]$'
test/e2e/framework/framework.go:152
Nov 11 05:18:50.094: Couldn't delete ns: "events-6608": namespace events-6608 was not deleted with limit: timed out waiting for the condition, pods remaining: 1 (&errors.errorString{s:"namespace events-6608 was not deleted with limit: timed out waiting for the condition, pods remaining: 1"})
test/e2e/framework/framework.go:336
				
				Click to see stdout/stderrfrom junit_01.xml

Filter through log files | View test history on testgrid


Kubernetes e2e suite [sig-apps] DisruptionController evictions: maxUnavailable allow single eviction, percentage => should allow an eviction 14m32s

go run hack/e2e.go -v --test --test_args='--ginkgo.focus=Kubernetes\se2e\ssuite\s\[sig\-apps\]\sDisruptionController\sevictions\:\smaxUnavailable\sallow\ssingle\seviction\,\spercentage\s\=\>\sshould\sallow\san\seviction$'
test/e2e/framework/framework.go:152
Nov 11 05:22:53.345: Couldn't delete ns: "disruption-5589": namespace disruption-5589 was not deleted with limit: timed out waiting for the condition, pods remaining: 3 (&errors.errorString{s:"namespace disruption-5589 was not deleted with limit: timed out waiting for the condition, pods remaining: 3"})
test/e2e/framework/framework.go:336
				
				Click to see stdout/stderrfrom junit_13.xml

Filter through log files | View test history on testgrid


Kubernetes e2e suite [sig-apps] DisruptionController evictions: maxUnavailable allow single eviction, percentage => should allow an eviction 16m1s

go run hack/e2e.go -v --test --test_args='--ginkgo.focus=Kubernetes\se2e\ssuite\s\[sig\-apps\]\sDisruptionController\sevictions\:\smaxUnavailable\sallow\ssingle\seviction\,\spercentage\s\=\>\sshould\sallow\san\seviction$'
test/e2e/framework/framework.go:152
Nov 11 05:38:54.837: Couldn't delete ns: "disruption-6298": namespace disruption-6298 was not deleted with limit: timed out waiting for the condition, pods remaining: 3 (&errors.errorString{s:"namespace disruption-6298 was not deleted with limit: timed out waiting for the condition, pods remaining: 3"})
test/e2e/framework/framework.go:336
				
				Click to see stdout/stderrfrom junit_13.xml

Filter through log files | View test history on testgrid


Kubernetes e2e suite [sig-apps] DisruptionController evictions: no PDB => should allow an eviction 13m14s

go run hack/e2e.go -v --test --test_args='--ginkgo.focus=Kubernetes\se2e\ssuite\s\[sig\-apps\]\sDisruptionController\sevictions\:\sno\sPDB\s\=\>\sshould\sallow\san\seviction$'
test/e2e/framework/framework.go:152
Nov 11 05:18:48.364: Couldn't delete ns: "disruption-8407": namespace disruption-8407 was not deleted with limit: timed out waiting for the condition, pods remaining: 1 (&errors.errorString{s:"namespace disruption-8407 was not deleted with limit: timed out waiting for the condition, pods remaining: 1"})
test/e2e/framework/framework.go:336
				
				Click to see stdout/stderrfrom junit_23.xml

Filter through log files | View test history on testgrid


Kubernetes e2e suite [sig-apps] StatefulSet [k8s.io] Basic StatefulSet functionality [StatefulSetBasic] should perform canary updates and phased rolling updates of template modifications [Conformance] 24m48s

go run hack/e2e.go -v --test --test_args='--ginkgo.focus=Kubernetes\se2e\ssuite\s\[sig\-apps\]\sStatefulSet\s\[k8s\.io\]\sBasic\sStatefulSet\sfunctionality\s\[StatefulSetBasic\]\sshould\sperform\scanary\supdates\sand\sphased\srolling\supdates\sof\stemplate\smodifications\s\[Conformance\]$'
test/e2e/apps/statefulset.go:88
Nov 11 05:21:49.009: Unexpected error:
    <*errors.errorString | 0xc0022123f0>: {
        s: "Failed to scale statefulset to 0 in 10m0s. Remaining pods:\n[ss2-1: deletion 2019-11-11 05:07:01 +0000 UTC, phase Running, readiness false]",
    }
    Failed to scale statefulset to 0 in 10m0s. Remaining pods:
    [ss2-1: deletion 2019-11-11 05:07:01 +0000 UTC, phase Running, readiness false]
occurred
test/e2e/framework/statefulset/rest.go:148
				
				Click to see stdout/stderrfrom junit_16.xml

Filter through log files | View test history on testgrid


Kubernetes e2e suite [sig-apps] StatefulSet [k8s.io] Basic StatefulSet functionality [StatefulSetBasic] should perform canary updates and phased rolling updates of template modifications [Conformance] 25m26s

go run hack/e2e.go -v --test --test_args='--ginkgo.focus=Kubernetes\se2e\ssuite\s\[sig\-apps\]\sStatefulSet\s\[k8s\.io\]\sBasic\sStatefulSet\sfunctionality\s\[StatefulSetBasic\]\sshould\sperform\scanary\supdates\sand\sphased\srolling\supdates\sof\stemplate\smodifications\s\[Conformance\]$'
test/e2e/apps/statefulset.go:88
Nov 11 05:47:13.510: Unexpected error:
    <*errors.errorString | 0xc001ab2790>: {
        s: "Failed to scale statefulset to 0 in 10m0s. Remaining pods:\n[ss2-1: deletion 2019-11-11 05:29:47 +0000 UTC, phase Running, readiness false]",
    }
    Failed to scale statefulset to 0 in 10m0s. Remaining pods:
    [ss2-1: deletion 2019-11-11 05:29:47 +0000 UTC, phase Running, readiness false]
occurred
test/e2e/framework/statefulset/rest.go:148
				
				Click to see stdout/stderrfrom junit_16.xml

Filter through log files | View test history on testgrid


Kubernetes e2e suite [sig-cli] Kubectl client Guestbook application should create and stop a working application [Conformance] 12m55s

go run hack/e2e.go -v --test --test_args='--ginkgo.focus=Kubernetes\se2e\ssuite\s\[sig\-cli\]\sKubectl\sclient\sGuestbook\sapplication\sshould\screate\sand\sstop\sa\sworking\sapplication\s\s\[Conformance\]$'
test/e2e/framework/framework.go:152
Nov 11 05:33:17.646: Couldn't delete ns: "kubectl-1290": namespace kubectl-1290 was not deleted with limit: timed out waiting for the condition, pods remaining: 2 (&errors.errorString{s:"namespace kubectl-1290 was not deleted with limit: timed out waiting for the condition, pods remaining: 2"})
test/e2e/framework/framework.go:336
				
				Click to see stdout/stderrfrom junit_07.xml

Filter through log files | View test history on testgrid


Kubernetes e2e suite [sig-cli] Kubectl client Update Demo should scale a replication controller [Conformance] 17m13s

go run hack/e2e.go -v --test --test_args='--ginkgo.focus=Kubernetes\se2e\ssuite\s\[sig\-cli\]\sKubectl\sclient\sUpdate\sDemo\sshould\sscale\sa\sreplication\scontroller\s\s\[Conformance\]$'
test/e2e/framework/framework.go:698
Nov 11 05:44:08.307: Timed out after 300 seconds waiting for name=update-demo pods to reach valid state
test/e2e/framework/rc_util.go:260
				
				Click to see stdout/stderrfrom junit_04.xml

Filter through log files | View test history on testgrid


Kubernetes e2e suite [sig-cli] Kubectl client Update Demo should scale a replication controller [Conformance] 20m21s

go run hack/e2e.go -v --test --test_args='--ginkgo.focus=Kubernetes\se2e\ssuite\s\[sig\-cli\]\sKubectl\sclient\sUpdate\sDemo\sshould\sscale\sa\sreplication\scontroller\s\s\[Conformance\]$'
test/e2e/framework/framework.go:698
Nov 11 05:25:52.820: Timed out after 300 seconds waiting for name=update-demo pods to reach valid state
test/e2e/framework/rc_util.go:260
				
				Click to see stdout/stderrfrom junit_04.xml

Filter through log files | View test history on testgrid


Kubernetes e2e suite [sig-network] Networking Granular Checks: Services should function for endpoint-Service: http 15m15s

go run hack/e2e.go -v --test --test_args='--ginkgo.focus=Kubernetes\se2e\ssuite\s\[sig\-network\]\sNetworking\sGranular\sChecks\:\sServices\sshould\sfunction\sfor\sendpoint\-Service\:\shttp$'
test/e2e/network/networking.go:141
Nov 11 05:22:50.196: Unexpected error:
    <*errors.errorString | 0xc0000d5090>: {
        s: "timed out waiting for the condition",
    }
    timed out waiting for the condition
occurred
test/e2e/framework/networking_utils.go:660
				
				Click to see stdout/stderrfrom junit_19.xml

Filter through log files | View test history on testgrid


Kubernetes e2e suite [sig-network] Services should be able to switch session affinity for service with type clusterIP 4m32s

go run hack/e2e.go -v --test --test_args='--ginkgo.focus=Kubernetes\se2e\ssuite\s\[sig\-network\]\sServices\sshould\sbe\sable\sto\sswitch\ssession\saffinity\sfor\sservice\swith\stype\sclusterIP$'
test/e2e/network/service.go:1807
Nov 11 05:00:07.512: Connection to 10.0.45.215:80 timed out or not enough responses.
test/e2e/framework/service/affinity_checker.go:55
				
				Click to see stdout/stderrfrom junit_02.xml

Filter through log files | View test history on testgrid


Kubernetes e2e suite [sig-scheduling] PreemptionExecutionPath runs ReplicaSets to verify preemption running path 4m29s

go run hack/e2e.go -v --test --test_args='--ginkgo.focus=Kubernetes\se2e\ssuite\s\[sig\-scheduling\]\sPreemptionExecutionPath\sruns\sReplicaSets\sto\sverify\spreemption\srunning\spath$'
test/e2e/scheduling/preemption.go:345
Nov 11 04:55:28.616: Unexpected error:
    <*errors.errorString | 0xc0019afb50>: {
        s: "replicaset \"rs-pod1\" never had desired number of .status.availableReplicas",
    }
    replicaset "rs-pod1" never had desired number of .status.availableReplicas
occurred
test/e2e/scheduling/preemption.go:510