This job view page is being replaced by Spyglass soon. Check out the new job view.
ResultFAILURE
Tests 6 failed / 608 succeeded
Started2019-05-23 00:40
Elapsed1h29m
Revision
Buildergke-prow-containerd-pool-99179761-0cmt
links{u'resultstore': {u'url': u'https://source.cloud.google.com/results/invocations/1c6272ee-ad43-4807-80da-fbeeb46f812a/targets/test'}}
pod455774a0-7cf3-11e9-99c9-36555a095607
resultstorehttps://source.cloud.google.com/results/invocations/1c6272ee-ad43-4807-80da-fbeeb46f812a/targets/test
infra-commite85087e74
job-versionv1.12.9-beta.0.48+3e39ad05dbde34
master_os_image
node_os_imagecos-u-73-11647-182-0
pod455774a0-7cf3-11e9-99c9-36555a095607
revisionv1.12.9-beta.0.48+3e39ad05dbde34

Test Failures


Cluster upgrade apparmor-upgrade 24m35s

go run hack/e2e.go -v --test --test_args='--ginkgo.focus=Cluster\supgrade\sapparmor\-upgrade$'
Should be able to get pod
Unexpected error:
    <*errors.StatusError | 0xc0020de990>: {
        ErrStatus: {
            TypeMeta: {Kind: "", APIVersion: ""},
            ListMeta: {SelfLink: "", ResourceVersion: "", Continue: ""},
            Status: "Failure",
            Message: "pods \"test-apparmor-mhfp6\" not found",
            Reason: "NotFound",
            Details: {
                Name: "test-apparmor-mhfp6",
                Group: "",
                Kind: "pods",
                UID: "",
                Causes: nil,
                RetryAfterSeconds: 0,
            },
            Code: 404,
        },
    }
    pods "test-apparmor-mhfp6" not found
occurred

k8s.io/kubernetes/test/e2e/upgrades.(*AppArmorUpgradeTest).verifyPodStillUp(0x89dfd58, 0xc000b8f400)
	/go/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/upgrades/apparmor.go:88 +0x159
k8s.io/kubernetes/test/e2e/upgrades.(*AppArmorUpgradeTest).Test(0x89dfd58, 0xc000b8f400, 0xc001c80420, 0x2)
	/go/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/upgrades/apparmor.go:72 +0x5a
k8s.io/kubernetes/test/e2e/lifecycle.(*chaosMonkeyAdapter).Test(0xc002253140, 0xc0020633e0)
	/go/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/lifecycle/cluster_upgrade.go:454 +0x309
k8s.io/kubernetes/test/e2e/chaosmonkey.(*chaosmonkey).Do.func1(0xc0020633e0, 0xc00192a850)
	/go/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/chaosmonkey/chaosmonkey.go:89 +0x76
created by k8s.io/kubernetes/test/e2e/chaosmonkey.(*chaosmonkey).Do
	/go/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/chaosmonkey/chaosmonkey.go:86 +0xa7
				from junit_upgradeupgrades.xml

Filter through log files | View test history on testgrid


Test 33m40s

error during ./hack/ginkgo-e2e.sh --ginkgo.skip=\[Slow\]|\[Serial\]|\[Disruptive\]|\[Flaky\]|\[Feature:.+\] --kubectl-path=../../../../kubernetes_skew/cluster/kubectl.sh --minStartupPods=8 --num-nodes=3 --report-dir=/workspace/_artifacts --disable-log-dump=true: exit status 1
				from junit_runner.xml

Filter through log files | View test history on testgrid


UpgradeTest 39m12s

error during kubetest --test --test_args=--ginkgo.focus=\[Feature:ClusterUpgrade\] --upgrade-image=gci --upgrade-target=ci/k8s-stable1 --num-nodes=3 --report-dir=/workspace/_artifacts --disable-log-dump=true --report-prefix=upgrade --check-version-skew=false: exit status 1
				from junit_runner.xml

Filter through log files | View test history on testgrid


[sig-apps] CronJob should remove from active list jobs that have been deleted 6m38s

go run hack/e2e.go -v --test --test_args='--ginkgo.focus=\[sig\-apps\]\sCronJob\sshould\sremove\sfrom\sactive\slist\sjobs\sthat\shave\sbeen\sdeleted$'
/go/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/apps/cronjob.go:192
Expected error:
    <*errors.errorString | 0xc4200d96b0>: {
        s: "timed out waiting for the condition",
    }
    timed out waiting for the condition
not to have occurred
/go/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/apps/cronjob.go:219
				
				Click to see stdout/stderrfrom junit_21.xml

Filter through log files | View test history on testgrid


[sig-cluster-lifecycle] Upgrade [Feature:Upgrade] cluster upgrade should maintain a functioning cluster [Feature:ClusterUpgrade] 38m54s

go run hack/e2e.go -v --test --test_args='--ginkgo.focus=\[sig\-cluster\-lifecycle\]\sUpgrade\s\[Feature\:Upgrade\]\scluster\supgrade\sshould\smaintain\sa\sfunctioning\scluster\s\[Feature\:ClusterUpgrade\]$'
/go/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/lifecycle/cluster_upgrade.go:143
Should be able to get pod
Unexpected error:
    <*errors.StatusError | 0xc0020de990>: {
        ErrStatus: {
            TypeMeta: {Kind: "", APIVersion: ""},
            ListMeta: {SelfLink: "", ResourceVersion: "", Continue: ""},
            Status: "Failure",
            Message: "pods \"test-apparmor-mhfp6\" not found",
            Reason: "NotFound",
            Details: {
                Name: "test-apparmor-mhfp6",
                Group: "",
                Kind: "pods",
                UID: "",
                Causes: nil,
                RetryAfterSeconds: 0,
            },
            Code: 404,
        },
    }
    pods "test-apparmor-mhfp6" not found
occurred
/go/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/upgrades/apparmor.go:88
				
				Click to see stdout/stderrfrom junit_upgrade01.xml

Filter through log files | View test history on testgrid


[sig-storage] In-tree Volumes [Driver: gluster] [Testpattern: Pre-provisioned PV (default fs)] subPath should support file as subpath 1m4s

go run hack/e2e.go -v --test --test_args='--ginkgo.focus=\[sig\-storage\]\sIn\-tree\sVolumes\s\[Driver\:\sgluster\]\s\[Testpattern\:\sPre\-provisioned\sPV\s\(default\sfs\)\]\ssubPath\sshould\ssupport\sfile\sas\ssubpath$'
/go/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/storage/testsuites/subpath.go:106
Expected error:
    <*errors.StatusError | 0xc420ca4fc0>: {
        ErrStatus: {
            TypeMeta: {Kind: "", APIVersion: ""},
            ListMeta: {SelfLink: "", ResourceVersion: "", Continue: ""},
            Status: "Failure",
            Message: "an error on the server (\"Internal Server Error: \\\"/api/v1/namespaces/e2e-tests-volumes-dqp5g/pods/gluster-server\\\": the server is currently unable to handle the request\") has prevented the request from succeeding (get pods gluster-server)",
            Reason: "InternalError",
            Details: {
                Name: "gluster-server",
                Group: "",
                Kind: "pods",
                UID: "",
                Causes: [
                    {
                        Type: "UnexpectedServerResponse",
                        Message: "Internal Server Error: \"/api/v1/namespaces/e2e-tests-volumes-dqp5g/pods/gluster-server\": the server is currently unable to handle the request",
                        Field: "",
                    },
                ],
                RetryAfterSeconds: 0,
            },
            Code: 500,
        },
    }
    an error on the server ("Internal Server Error: \"/api/v1/namespaces/e2e-tests-volumes-dqp5g/pods/gluster-server\": the server is currently unable to handle the request") has prevented the request from succeeding (get pods gluster-server)
not to have occurred
/go/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/volume_util.go:344
				
				Click to see stdout/stderrfrom junit_13.xml

Filter through log files | View test history on testgrid


Show 608 Passed Tests