This job view page is being replaced by Spyglass soon. Check out the new job view.
ResultFAILURE
Tests 12 failed / 136 succeeded
Started2019-05-22 12:32
Elapsed10h49m
Revision
Buildergke-prow-containerd-pool-99179761-9sg5
links{u'resultstore': {u'url': u'https://source.cloud.google.com/results/invocations/20548231-1a4d-43bd-ae71-f0f5a13097c1/targets/test'}}
pod7e8d9696-7c8d-11e9-a44f-96e6949cb91f
resultstorehttps://source.cloud.google.com/results/invocations/20548231-1a4d-43bd-ae71-f0f5a13097c1/targets/test
infra-commite96e1ab77
job-versionv1.13.7-beta.0.19+b3866462b51216
master_os_image
node_os_imageubuntu-gke-1804-d1703-0-v20190517
pod7e8d9696-7c8d-11e9-a44f-96e6949cb91f
revisionv1.13.7-beta.0.19+b3866462b51216

Test Failures


Test 10h33m

error during ./hack/ginkgo-e2e.sh --ginkgo.focus=\[Serial\]|\[Disruptive\] --ginkgo.skip=\[Flaky\]|\[Feature:.+\] --minStartupPods=8 --num-nodes=3 --report-dir=/workspace/_artifacts --disable-log-dump=true: exit status 1
				from junit_runner.xml

Filter through log files | View test history on testgrid


[k8s.io] EquivalenceCache [Serial] validates pod affinity works properly when new replica pod is scheduled 1h1m

go run hack/e2e.go -v --test --test_args='--ginkgo.focus=EquivalenceCache\s\[Serial\]\svalidates\spod\saffinity\sworks\sproperly\swhen\snew\sreplica\spod\sis\sscheduled$'
/go/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/scheduling/equivalence_cache_predicates.go:50
Expected error:
    <*errors.errorString | 0xc0006f6cd0>: {
        s: "Waiting for terminating namespaces to be deleted timed out",
    }
    Waiting for terminating namespaces to be deleted timed out
not to have occurred
/go/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/scheduling/equivalence_cache_predicates.go:57
				
				Click to see stdout/stderrfrom junit_01.xml

Filter through log files | View test history on testgrid


[sig-autoscaling] [HPA] Horizontal pod autoscaling (scale resource: CPU) [sig-autoscaling] [Serial] [Slow] Deployment Should scale from 1 pod to 3 pods and from 3 to 5 12m23s

go run hack/e2e.go -v --test --test_args='--ginkgo.focus=\[sig\-autoscaling\]\s\[HPA\]\sHorizontal\spod\sautoscaling\s\(scale\sresource\:\sCPU\)\s\[sig\-autoscaling\]\s\[Serial\]\s\[Slow\]\sDeployment\sShould\sscale\sfrom\s1\spod\sto\s3\spods\sand\sfrom\s3\sto\s5$'
/go/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/autoscaling/horizontal_pod_autoscaling.go:40
Expected error:
    <*errors.errorString | 0xc0018c4310>: {
        s: "Only 0 pods started out of 1",
    }
    Only 0 pods started out of 1
not to have occurred
/go/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/common/autoscaling_utils.go:482
				
				Click to see stdout/stderrfrom junit_01.xml

Filter through log files | View test history on testgrid


[sig-cluster-lifecycle] Nodes [Disruptive] Resize [Slow] should be able to delete nodes 18m59s

go run hack/e2e.go -v --test --test_args='--ginkgo.focus=\[sig\-cluster\-lifecycle\]\sNodes\s\[Disruptive\]\sResize\s\[Slow\]\sshould\sbe\sable\sto\sdelete\snodes$'
/go/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/lifecycle/resize_nodes.go:110
Expected error:
    <*errors.errorString | 0xc0011c3a40>: {
        s: "failed to wait for pods running: [timed out waiting for the condition]",
    }
    failed to wait for pods running: [timed out waiting for the condition]
not to have occurred
/go/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/lifecycle/resize_nodes.go:136
				
				Click to see stdout/stderrfrom junit_01.xml

Filter through log files | View test history on testgrid


[sig-scheduling] PodPriorityResolution [Serial] validates critical system priorities are created and resolved 1h0m

go run hack/e2e.go -v --test --test_args='--ginkgo.focus=\[sig\-scheduling\]\sPodPriorityResolution\s\[Serial\]\svalidates\scritical\ssystem\spriorities\sare\screated\sand\sresolved$'
/go/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/scheduling/preemption.go:325
Expected error:
    <*errors.errorString | 0xc00102e980>: {
        s: "Waiting for terminating namespaces to be deleted timed out",
    }
    Waiting for terminating namespaces to be deleted timed out
not to have occurred
/go/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/scheduling/preemption.go:330
				
				Click to see stdout/stderrfrom junit_01.xml

Filter through log files | View test history on testgrid


[sig-scheduling] TaintBasedEvictions [Serial] Checks that the node becomes unreachable 1m46s

go run hack/e2e.go -v --test --test_args='--ginkgo.focus=\[sig\-scheduling\]\sTaintBasedEvictions\s\[Serial\]\sChecks\sthat\sthe\snode\sbecomes\sunreachable$'
/go/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/scheduling/taint_based_evictions.go:75
Expected error:
    <*errors.errorString | 0xc0023aac00>: {
        s: "expect node gke-test-7671b0fd74-default-pool-8381440f-whbg to have taint = true within 30s: timed out waiting for the condition",
    }
    expect node gke-test-7671b0fd74-default-pool-8381440f-whbg to have taint = true within 30s: timed out waiting for the condition
not to have occurred
/go/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/scheduling/taint_based_evictions.go:162
				
				Click to see stdout/stderrfrom junit_01.xml

Filter through log files | View test history on testgrid


[sig-storage] CSI Volumes [Driver: pd.csi.storage.gke.io][Serial] [Testpattern: Dynamic PV (default fs)] subPath should fail if subpath directory is outside the volume [Slow] 11m8s

go run hack/e2e.go -v --test --test_args='--ginkgo.focus=\[sig\-storage\]\sCSI\sVolumes\s\[Driver\:\spd\.csi\.storage\.gke\.io\]\[Serial\]\s\[Testpattern\:\sDynamic\sPV\s\(default\sfs\)\]\ssubPath\sshould\sfail\sif\ssubpath\sdirectory\sis\soutside\sthe\svolume\s\[Slow\]$'
/go/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:154
May 22 20:52:42.356: Couldn't delete ns: "e2e-tests-csi-volumes-lsc4r": namespace e2e-tests-csi-volumes-lsc4r was not deleted with limit: timed out waiting for the condition, pods remaining: 1 (&errors.errorString{s:"namespace e2e-tests-csi-volumes-lsc4r was not deleted with limit: timed out waiting for the condition, pods remaining: 1"})
/go/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:345
				
				Click to see stdout/stderrfrom junit_01.xml

Filter through log files | View test history on testgrid


[sig-storage] CSI Volumes [Driver: pd.csi.storage.gke.io][Serial] [Testpattern: Dynamic PV (default fs)] subPath should support creating multiple subpath from same volumes [Slow] 12m36s

go run hack/e2e.go -v --test --test_args='--ginkgo.focus=\[sig\-storage\]\sCSI\sVolumes\s\[Driver\:\spd\.csi\.storage\.gke\.io\]\[Serial\]\s\[Testpattern\:\sDynamic\sPV\s\(default\sfs\)\]\ssubPath\sshould\ssupport\screating\smultiple\ssubpath\sfrom\ssame\svolumes\s\[Slow\]$'
/go/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:154
May 22 20:04:38.538: Couldn't delete ns: "e2e-tests-csi-volumes-jd5gp": namespace e2e-tests-csi-volumes-jd5gp was not deleted with limit: timed out waiting for the condition, pods remaining: 2 (&errors.errorString{s:"namespace e2e-tests-csi-volumes-jd5gp was not deleted with limit: timed out waiting for the condition, pods remaining: 2"})
/go/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:345
				
				Click to see stdout/stderrfrom junit_01.xml

Filter through log files | View test history on testgrid


[sig-storage] CSI Volumes [Driver: pd.csi.storage.gke.io][Serial] [Testpattern: Dynamic PV (ext3)] volumes should allow exec of files on the volume 11m21s

go run hack/e2e.go -v --test --test_args='--ginkgo.focus=\[sig\-storage\]\sCSI\sVolumes\s\[Driver\:\spd\.csi\.storage\.gke\.io\]\[Serial\]\s\[Testpattern\:\sDynamic\sPV\s\(ext3\)\]\svolumes\sshould\sallow\sexec\sof\sfiles\son\sthe\svolume$'
/go/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:154
May 22 23:13:23.571: Couldn't delete ns: "e2e-tests-csi-volumes-j5p2r": namespace e2e-tests-csi-volumes-j5p2r was not deleted with limit: timed out waiting for the condition, pods remaining: 1 (&errors.errorString{s:"namespace e2e-tests-csi-volumes-j5p2r was not deleted with limit: timed out waiting for the condition, pods remaining: 1"})
/go/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:345
				
				Click to see stdout/stderrfrom junit_01.xml

Filter through log files | View test history on testgrid


[sig-storage] CSI Volumes [Driver: pd.csi.storage.gke.io][Serial] [Testpattern: Dynamic PV (ext4)] volumes should be mountable 12m33s

go run hack/e2e.go -v --test --test_args='--ginkgo.focus=\[sig\-storage\]\sCSI\sVolumes\s\[Driver\:\spd\.csi\.storage\.gke\.io\]\[Serial\]\s\[Testpattern\:\sDynamic\sPV\s\(ext4\)\]\svolumes\sshould\sbe\smountable$'
/go/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:154
May 22 20:36:30.143: Couldn't delete ns: "e2e-tests-csi-volumes-vxr9m": namespace e2e-tests-csi-volumes-vxr9m was not deleted with limit: timed out waiting for the condition, pods remaining: 1 (&errors.errorString{s:"namespace e2e-tests-csi-volumes-vxr9m was not deleted with limit: timed out waiting for the condition, pods remaining: 1"})
/go/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:345
				
				Click to see stdout/stderrfrom junit_01.xml

Filter through log files | View test history on testgrid


[sig-storage] In-tree Volumes [Driver: nfs] [Testpattern: Dynamic PV (default fs)] subPath should unmount if pod is gracefully deleted while kubelet is down [Disruptive][Slow] 27m34s

go run hack/e2e.go -v --test --test_args='--ginkgo.focus=\[sig\-storage\]\sIn\-tree\sVolumes\s\[Driver\:\snfs\]\s\[Testpattern\:\sDynamic\sPV\s\(default\sfs\)\]\ssubPath\sshould\sunmount\sif\spod\sis\sgracefully\sdeleted\swhile\skubelet\sis\sdown\s\[Disruptive\]\[Slow\]$'
/go/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/storage/testsuites/subpath.go:313
Expected pod to be not found.
Expected error:
    <*errors.errorString | 0xc00009b860>: {
        s: "timed out waiting for the condition",
    }
    timed out waiting for the condition
not to have occurred
/go/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/storage/utils/utils.go:242
				
				Click to see stdout/stderrfrom junit_01.xml

Find to mentions in log files | View test history on testgrid


[sig-storage] PersistentVolumes-local Local volume provisioner [Serial] should create and recreate local persistent volume 5m30s

go run hack/e2e.go -v --test --test_args='--ginkgo.focus=\[sig\-storage\]\sPersistentVolumes\-local\s\sLocal\svolume\sprovisioner\s\[Serial\]\sshould\screate\sand\srecreate\slocal\spersistent\svolume$'
/go/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/storage/persistent_volumes-local.go:436
Expected error:
    <*errors.errorString | 0xc00009b860>: {
        s: "timed out waiting for the condition",
    }
    timed out waiting for the condition
not to have occurred
/go/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/storage/persistent_volumes-local.go:984
				
				Click to see stdout/stderrfrom junit_01.xml

Filter through log files | View test history on testgrid


Show 136 Passed Tests

Show 2032 Skipped Tests