This job view page is being replaced by Spyglass soon. Check out the new job view.
ResultFAILURE
Tests 44 failed / 14 succeeded
Started2020-10-26 20:39
Elapsed5h1m
Revision
Builder50472467-17cb-11eb-8cf5-bee87f5cdaa3
infra-commit3118d2e26
job-versionv1.20.0-alpha.3.144+6675eba3eff1c8-dirty
repok8s.io/kubernetes
repo-commit6675eba3eff1c8e565c4060a9c1396f75da7cc3e
repos{u'k8s.io/kubernetes': u'master'}
revisionv1.20.0-alpha.3.144+6675eba3eff1c8-dirty

Test Failures


E2eNode Suite [k8s.io] Container Manager Misc [Serial] Validate OOM score adjustments [NodeFeature:OOMScoreAdj] once the node is setup pod infra containers oom-score-adj should be -998 and best effort container's should be 1000 30s

go run hack/e2e.go -v --test --test_args='--ginkgo.focus=E2eNode\sSuite\s\[k8s\.io\]\sContainer\sManager\sMisc\s\[Serial\]\sValidate\sOOM\sscore\sadjustments\s\[NodeFeature\:OOMScoreAdj\]\sonce\sthe\snode\sis\ssetup\s\spod\sinfra\scontainers\soom\-score\-adj\sshould\sbe\s\-998\sand\sbest\seffort\scontainer\'s\sshould\sbe\s1000$'
/go/src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:174
Unexpected error:
    <*errors.errorString | 0xc0002329b0>: {
        s: "timed out waiting for the condition",
    }
    timed out waiting for the condition
occurred
/go/src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:236
				
				Click to see stdout/stderrfrom junit_cos-stable2_01.xml

Filter through log files | View test history on testgrid


E2eNode Suite [k8s.io] Container Manager Misc [Serial] Validate OOM score adjustments [NodeFeature:OOMScoreAdj] once the node is setup Kubelet's oom-score-adj should be -999 30s

go run hack/e2e.go -v --test --test_args='--ginkgo.focus=E2eNode\sSuite\s\[k8s\.io\]\sContainer\sManager\sMisc\s\[Serial\]\sValidate\sOOM\sscore\sadjustments\s\[NodeFeature\:OOMScoreAdj\]\sonce\sthe\snode\sis\ssetup\sKubelet\'s\soom\-score\-adj\sshould\sbe\s\-999$'
/go/src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:174
Unexpected error:
    <*errors.errorString | 0xc0002329b0>: {
        s: "timed out waiting for the condition",
    }
    timed out waiting for the condition
occurred
/go/src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:236
				
				Click to see stdout/stderrfrom junit_cos-stable2_01.xml

Filter through log files | View test history on testgrid


E2eNode Suite [k8s.io] Container Manager Misc [Serial] Validate OOM score adjustments [NodeFeature:OOMScoreAdj] once the node is setup burstable container's oom-score-adj should be between [2, 1000) 30s

go run hack/e2e.go -v --test --test_args='--ginkgo.focus=E2eNode\sSuite\s\[k8s\.io\]\sContainer\sManager\sMisc\s\[Serial\]\sValidate\sOOM\sscore\sadjustments\s\[NodeFeature\:OOMScoreAdj\]\sonce\sthe\snode\sis\ssetup\sburstable\scontainer\'s\soom\-score\-adj\sshould\sbe\sbetween\s\[2\,\s1000\)$'
/go/src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:174
Unexpected error:
    <*errors.errorString | 0xc0002329b0>: {
        s: "timed out waiting for the condition",
    }
    timed out waiting for the condition
occurred
/go/src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:236
				
				Click to see stdout/stderrfrom junit_cos-stable2_01.xml

Filter through log files | View test history on testgrid


E2eNode Suite [k8s.io] Container Manager Misc [Serial] Validate OOM score adjustments [NodeFeature:OOMScoreAdj] once the node is setup container runtime's oom-score-adj should be -999 30s

go run hack/e2e.go -v --test --test_args='--ginkgo.focus=E2eNode\sSuite\s\[k8s\.io\]\sContainer\sManager\sMisc\s\[Serial\]\sValidate\sOOM\sscore\sadjustments\s\[NodeFeature\:OOMScoreAdj\]\sonce\sthe\snode\sis\ssetup\scontainer\sruntime\'s\soom\-score\-adj\sshould\sbe\s\-999$'
/go/src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:174
Unexpected error:
    <*errors.errorString | 0xc0002329b0>: {
        s: "timed out waiting for the condition",
    }
    timed out waiting for the condition
occurred
/go/src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:236
				
				Click to see stdout/stderrfrom junit_cos-stable2_01.xml

Filter through log files | View test history on testgrid


E2eNode Suite [k8s.io] Container Manager Misc [Serial] Validate OOM score adjustments [NodeFeature:OOMScoreAdj] once the node is setup guaranteed container's oom-score-adj should be -998 30s

go run hack/e2e.go -v --test --test_args='--ginkgo.focus=E2eNode\sSuite\s\[k8s\.io\]\sContainer\sManager\sMisc\s\[Serial\]\sValidate\sOOM\sscore\sadjustments\s\[NodeFeature\:OOMScoreAdj\]\sonce\sthe\snode\sis\ssetup\sguaranteed\scontainer\'s\soom\-score\-adj\sshould\sbe\s\-998$'
/go/src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:174
Unexpected error:
    <*errors.errorString | 0xc0002329b0>: {
        s: "timed out waiting for the condition",
    }
    timed out waiting for the condition
occurred
/go/src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:236
				
				Click to see stdout/stderrfrom junit_cos-stable2_01.xml

Filter through log files | View test history on testgrid


E2eNode Suite [k8s.io] ContainerLogRotation [Slow] [Serial] [Disruptive] when a container generates a lot of log should be rotated and limited to a fixed amount of files 30s

go run hack/e2e.go -v --test --test_args='--ginkgo.focus=E2eNode\sSuite\s\[k8s\.io\]\sContainerLogRotation\s\[Slow\]\s\[Serial\]\s\[Disruptive\]\swhen\sa\scontainer\sgenerates\sa\slot\sof\slog\sshould\sbe\srotated\sand\slimited\sto\sa\sfixed\samount\sof\sfiles$'
/go/src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:174
Unexpected error:
    <*errors.errorString | 0xc0002329b0>: {
        s: "timed out waiting for the condition",
    }
    timed out waiting for the condition
occurred
/go/src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:236
				
				Click to see stdout/stderrfrom junit_cos-stable2_01.xml

Filter through log files | View test history on testgrid


E2eNode Suite [k8s.io] CriticalPod [Serial] [Disruptive] [NodeFeature:CriticalPod] when we need to admit a critical pod should be able to create and delete a critical pod 30s

go run hack/e2e.go -v --test --test_args='--ginkgo.focus=E2eNode\sSuite\s\[k8s\.io\]\sCriticalPod\s\[Serial\]\s\[Disruptive\]\s\[NodeFeature\:CriticalPod\]\swhen\swe\sneed\sto\sadmit\sa\scritical\spod\sshould\sbe\sable\sto\screate\sand\sdelete\sa\scritical\spod$'
/go/src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:174
Unexpected error:
    <*errors.errorString | 0xc0002329b0>: {
        s: "timed out waiting for the condition",
    }
    timed out waiting for the condition
occurred
/go/src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:236
				
				Click to see stdout/stderrfrom junit_cos-stable2_01.xml

Filter through log files | View test history on testgrid


E2eNode Suite [k8s.io] Density [Serial] [Slow] create a batch of pods latency/resource should be within limit when create 10 pods with 0s interval 30s

go run hack/e2e.go -v --test --test_args='--ginkgo.focus=E2eNode\sSuite\s\[k8s\.io\]\sDensity\s\[Serial\]\s\[Slow\]\screate\sa\sbatch\sof\spods\slatency\/resource\sshould\sbe\swithin\slimit\swhen\screate\s10\spods\swith\s0s\sinterval$'
/go/src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:174
Unexpected error:
    <*errors.errorString | 0xc0002329b0>: {
        s: "timed out waiting for the condition",
    }
    timed out waiting for the condition
occurred
/go/src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:236
				
				Click to see stdout/stderrfrom junit_cos-stable2_01.xml

Filter through log files | View test history on testgrid


E2eNode Suite [k8s.io] Density [Serial] [Slow] create a sequence of pods latency/resource should be within limit when create 10 pods with 50 background pods 30s

go run hack/e2e.go -v --test --test_args='--ginkgo.focus=E2eNode\sSuite\s\[k8s\.io\]\sDensity\s\[Serial\]\s\[Slow\]\screate\sa\ssequence\sof\spods\slatency\/resource\sshould\sbe\swithin\slimit\swhen\screate\s10\spods\swith\s50\sbackground\spods$'
/go/src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:174
Unexpected error:
    <*errors.errorString | 0xc0002329b0>: {
        s: "timed out waiting for the condition",
    }
    timed out waiting for the condition
occurred
/go/src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:236
				
				Click to see stdout/stderrfrom junit_cos-stable2_01.xml

Filter through log files | View test history on testgrid


E2eNode Suite [k8s.io] Device Plugin [Feature:DevicePluginProbe][NodeFeature:DevicePluginProbe][Serial] DevicePlugin Verifies the Kubelet device plugin functionality. 30s

go run hack/e2e.go -v --test --test_args='--ginkgo.focus=E2eNode\sSuite\s\[k8s\.io\]\sDevice\sPlugin\s\[Feature\:DevicePluginProbe\]\[NodeFeature\:DevicePluginProbe\]\[Serial\]\sDevicePlugin\sVerifies\sthe\sKubelet\sdevice\splugin\sfunctionality\.$'
/go/src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:174
Unexpected error:
    <*errors.errorString | 0xc0002329b0>: {
        s: "timed out waiting for the condition",
    }
    timed out waiting for the condition
occurred
/go/src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:236
				
				Click to see stdout/stderrfrom junit_cos-stable2_01.xml

Filter through log files | View test history on testgrid


E2eNode Suite [k8s.io] Downward API [Serial] [Disruptive] [NodeFeature:EphemeralStorage] Downward API tests for local ephemeral storage should provide default limits.ephemeral-storage from node allocatable 30s

go run hack/e2e.go -v --test --test_args='--ginkgo.focus=E2eNode\sSuite\s\[k8s\.io\]\sDownward\sAPI\s\[Serial\]\s\[Disruptive\]\s\[NodeFeature\:EphemeralStorage\]\sDownward\sAPI\stests\sfor\slocal\sephemeral\sstorage\sshould\sprovide\sdefault\slimits\.ephemeral\-storage\sfrom\snode\sallocatable$'
/go/src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:174
Unexpected error:
    <*errors.errorString | 0xc0002329b0>: {
        s: "timed out waiting for the condition",
    }
    timed out waiting for the condition
occurred
/go/src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:236
				
				Click to see stdout/stderrfrom junit_cos-stable2_01.xml

Filter through log files | View test history on testgrid


E2eNode Suite [k8s.io] GarbageCollect [Serial][NodeFeature:GarbageCollect] Garbage Collection Test: Many Pods with Many Restarting Containers Should eventually garbage collect containers when we exceed the number of dead containers per container 30s

go run hack/e2e.go -v --test --test_args='--ginkgo.focus=E2eNode\sSuite\s\[k8s\.io\]\sGarbageCollect\s\[Serial\]\[NodeFeature\:GarbageCollect\]\sGarbage\sCollection\sTest\:\sMany\sPods\swith\sMany\sRestarting\sContainers\sShould\seventually\sgarbage\scollect\scontainers\swhen\swe\sexceed\sthe\snumber\sof\sdead\scontainers\sper\scontainer$'
/go/src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:174
Unexpected error:
    <*errors.errorString | 0xc0002329b0>: {
        s: "timed out waiting for the condition",
    }
    timed out waiting for the condition
occurred
/go/src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:236
				
				Click to see stdout/stderrfrom junit_cos-stable2_01.xml

Filter through log files | View test history on testgrid


E2eNode Suite [k8s.io] GarbageCollect [Serial][NodeFeature:GarbageCollect] Garbage Collection Test: Many Restarting Containers Should eventually garbage collect containers when we exceed the number of dead containers per container 30s

go run hack/e2e.go -v --test --test_args='--ginkgo.focus=E2eNode\sSuite\s\[k8s\.io\]\sGarbageCollect\s\[Serial\]\[NodeFeature\:GarbageCollect\]\sGarbage\sCollection\sTest\:\sMany\sRestarting\sContainers\sShould\seventually\sgarbage\scollect\scontainers\swhen\swe\sexceed\sthe\snumber\sof\sdead\scontainers\sper\scontainer$'
/go/src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:174
Unexpected error:
    <*errors.errorString | 0xc0002329b0>: {
        s: "timed out waiting for the condition",
    }
    timed out waiting for the condition
occurred
/go/src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:236
				
				Click to see stdout/stderrfrom junit_cos-stable2_01.xml

Filter through log files | View test history on testgrid


E2eNode Suite [k8s.io] GarbageCollect [Serial][NodeFeature:GarbageCollect] Garbage Collection Test: One Non-restarting Container Should eventually garbage collect containers when we exceed the number of dead containers per container 30s

go run hack/e2e.go -v --test --test_args='--ginkgo.focus=E2eNode\sSuite\s\[k8s\.io\]\sGarbageCollect\s\[Serial\]\[NodeFeature\:GarbageCollect\]\sGarbage\sCollection\sTest\:\sOne\sNon\-restarting\sContainer\sShould\seventually\sgarbage\scollect\scontainers\swhen\swe\sexceed\sthe\snumber\sof\sdead\scontainers\sper\scontainer$'
/go/src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:174
Unexpected error:
    <*errors.errorString | 0xc0002329b0>: {
        s: "timed out waiting for the condition",
    }
    timed out waiting for the condition
occurred
/go/src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:236
				
				Click to see stdout/stderrfrom junit_cos-stable2_01.xml

Filter through log files | View test history on testgrid


E2eNode Suite [k8s.io] InodeEviction [Slow] [Serial] [Disruptive][NodeFeature:Eviction] when we run containers that should cause DiskPressure should eventually evict all of the correct pods 30s

go run hack/e2e.go -v --test --test_args='--ginkgo.focus=E2eNode\sSuite\s\[k8s\.io\]\sInodeEviction\s\[Slow\]\s\[Serial\]\s\[Disruptive\]\[NodeFeature\:Eviction\]\swhen\swe\srun\scontainers\sthat\sshould\scause\sDiskPressure\s\sshould\seventually\sevict\sall\sof\sthe\scorrect\spods$'
/go/src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:174
Unexpected error:
    <*errors.errorString | 0xc0002329b0>: {
        s: "timed out waiting for the condition",
    }
    timed out waiting for the condition
occurred
/go/src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:236
				
				Click to see stdout/stderrfrom junit_cos-stable2_01.xml

Filter through log files | View test history on testgrid


E2eNode Suite [k8s.io] LocalStorageCapacityIsolationQuotaMonitoring [Slow] [Serial] [Disruptive] [Feature:LocalStorageCapacityIsolationQuota][NodeFeature:LSCIQuotaMonitoring] when we run containers that should cause use quotas for LSCI monitoring (quotas enabled: false) should eventually evict all of the correct pods 30s

go run hack/e2e.go -v --test --test_args='--ginkgo.focus=E2eNode\sSuite\s\[k8s\.io\]\sLocalStorageCapacityIsolationQuotaMonitoring\s\[Slow\]\s\[Serial\]\s\[Disruptive\]\s\[Feature\:LocalStorageCapacityIsolationQuota\]\[NodeFeature\:LSCIQuotaMonitoring\]\swhen\swe\srun\scontainers\sthat\sshould\scause\suse\squotas\sfor\sLSCI\smonitoring\s\(quotas\senabled\:\sfalse\)\s\sshould\seventually\sevict\sall\sof\sthe\scorrect\spods$'
/go/src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:174
Unexpected error:
    <*errors.errorString | 0xc0002329b0>: {
        s: "timed out waiting for the condition",
    }
    timed out waiting for the condition
occurred
/go/src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:236
				
				Click to see stdout/stderrfrom junit_cos-stable2_01.xml

Filter through log files | View test history on testgrid


E2eNode Suite [k8s.io] LocalStorageCapacityIsolationQuotaMonitoring [Slow] [Serial] [Disruptive] [Feature:LocalStorageCapacityIsolationQuota][NodeFeature:LSCIQuotaMonitoring] when we run containers that should cause use quotas for LSCI monitoring (quotas enabled: true) should eventually evict all of the correct pods 30s

go run hack/e2e.go -v --test --test_args='--ginkgo.focus=E2eNode\sSuite\s\[k8s\.io\]\sLocalStorageCapacityIsolationQuotaMonitoring\s\[Slow\]\s\[Serial\]\s\[Disruptive\]\s\[Feature\:LocalStorageCapacityIsolationQuota\]\[NodeFeature\:LSCIQuotaMonitoring\]\swhen\swe\srun\scontainers\sthat\sshould\scause\suse\squotas\sfor\sLSCI\smonitoring\s\(quotas\senabled\:\strue\)\s\sshould\seventually\sevict\sall\sof\sthe\scorrect\spods$'
/go/src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:174
Unexpected error:
    <*errors.errorString | 0xc0002329b0>: {
        s: "timed out waiting for the condition",
    }
    timed out waiting for the condition
occurred
/go/src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:236
				
				Click to see stdout/stderrfrom junit_cos-stable2_01.xml

Filter through log files | View test history on testgrid


E2eNode Suite [k8s.io] LocalStorageEviction [Slow] [Serial] [Disruptive][NodeFeature:Eviction] when we run containers that should cause DiskPressure should eventually evict all of the correct pods 30s

go run hack/e2e.go -v --test --test_args='--ginkgo.focus=E2eNode\sSuite\s\[k8s\.io\]\sLocalStorageEviction\s\[Slow\]\s\[Serial\]\s\[Disruptive\]\[NodeFeature\:Eviction\]\swhen\swe\srun\scontainers\sthat\sshould\scause\sDiskPressure\s\sshould\seventually\sevict\sall\sof\sthe\scorrect\spods$'
/go/src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:174
Unexpected error:
    <*errors.errorString | 0xc0002329b0>: {
        s: "timed out waiting for the condition",
    }
    timed out waiting for the condition
occurred
/go/src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:236
				
				Click to see stdout/stderrfrom junit_cos-stable2_01.xml

Filter through log files | View test history on testgrid


E2eNode Suite [k8s.io] LocalStorageSoftEviction [Slow] [Serial] [Disruptive][NodeFeature:Eviction] when we run containers that should cause DiskPressure should eventually evict all of the correct pods 30s

go run hack/e2e.go -v --test --test_args='--ginkgo.focus=E2eNode\sSuite\s\[k8s\.io\]\sLocalStorageSoftEviction\s\[Slow\]\s\[Serial\]\s\[Disruptive\]\[NodeFeature\:Eviction\]\swhen\swe\srun\scontainers\sthat\sshould\scause\sDiskPressure\s\sshould\seventually\sevict\sall\sof\sthe\scorrect\spods$'
/go/src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:174
Unexpected error:
    <*errors.errorString | 0xc0002329b0>: {
        s: "timed out waiting for the condition",
    }
    timed out waiting for the condition
occurred
/go/src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:236
				
				Click to see stdout/stderrfrom junit_cos-stable2_01.xml

Filter through log files | View test history on testgrid


E2eNode Suite [k8s.io] MemoryAllocatableEviction [Slow] [Serial] [Disruptive][NodeFeature:Eviction] when we run containers that should cause MemoryPressure should eventually evict all of the correct pods 30s

go run hack/e2e.go -v --test --test_args='--ginkgo.focus=E2eNode\sSuite\s\[k8s\.io\]\sMemoryAllocatableEviction\s\[Slow\]\s\[Serial\]\s\[Disruptive\]\[NodeFeature\:Eviction\]\swhen\swe\srun\scontainers\sthat\sshould\scause\sMemoryPressure\s\sshould\seventually\sevict\sall\sof\sthe\scorrect\spods$'
/go/src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:174
Unexpected error:
    <*errors.errorString | 0xc0002329b0>: {
        s: "timed out waiting for the condition",
    }
    timed out waiting for the condition
occurred
/go/src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:236
				
				Click to see stdout/stderrfrom junit_cos-stable2_01.xml

Filter through log files | View test history on testgrid


E2eNode Suite [k8s.io] NVIDIA GPU Device Plugin [Feature:GPUDevicePlugin][NodeFeature:GPUDevicePlugin][Serial] [Disruptive] DevicePlugin checks that when Kubelet restarts exclusive GPU assignation to pods is kept. 30s

go run hack/e2e.go -v --test --test_args='--ginkgo.focus=E2eNode\sSuite\s\[k8s\.io\]\sNVIDIA\sGPU\sDevice\sPlugin\s\[Feature\:GPUDevicePlugin\]\[NodeFeature\:GPUDevicePlugin\]\[Serial\]\s\[Disruptive\]\sDevicePlugin\schecks\sthat\swhen\sKubelet\srestarts\sexclusive\sGPU\sassignation\sto\spods\sis\skept\.$'
/go/src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:174
Unexpected error:
    <*errors.errorString | 0xc0002329b0>: {
        s: "timed out waiting for the condition",
    }
    timed out waiting for the condition
occurred
/go/src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:236
				
				Click to see stdout/stderrfrom junit_cos-stable2_01.xml

Filter through log files | View test history on testgrid


E2eNode Suite [k8s.io] Node Container Manager [Serial] Validate Node Allocatable [NodeFeature:NodeAllocatable] sets up the node and runs the test 30s

go run hack/e2e.go -v --test --test_args='--ginkgo.focus=E2eNode\sSuite\s\[k8s\.io\]\sNode\sContainer\sManager\s\[Serial\]\sValidate\sNode\sAllocatable\s\[NodeFeature\:NodeAllocatable\]\ssets\sup\sthe\snode\sand\sruns\sthe\stest$'
/go/src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:174
Unexpected error:
    <*errors.errorString | 0xc0002329b0>: {
        s: "timed out waiting for the condition",
    }
    timed out waiting for the condition
occurred
/go/src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:236
				
				Click to see stdout/stderrfrom junit_cos-stable2_01.xml

Filter through log files | View test history on testgrid


E2eNode Suite [k8s.io] NodeProblemDetector [NodeFeature:NodeProblemDetector] [Serial] [k8s.io] SystemLogMonitor should generate node condition and events for corresponding errors 30s

go run hack/e2e.go -v --test --test_args='--ginkgo.focus=E2eNode\sSuite\s\[k8s\.io\]\sNodeProblemDetector\s\[NodeFeature\:NodeProblemDetector\]\s\[Serial\]\s\[k8s\.io\]\sSystemLogMonitor\sshould\sgenerate\snode\scondition\sand\sevents\sfor\scorresponding\serrors$'
/go/src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:174
Unexpected error:
    <*errors.errorString | 0xc0002329b0>: {
        s: "timed out waiting for the condition",
    }
    timed out waiting for the condition
occurred
/go/src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:236
				
				Click to see stdout/stderrfrom junit_cos-stable2_01.xml

Filter through log files | View test history on testgrid


E2eNode Suite [k8s.io] PriorityLocalStorageEvictionOrdering [Slow] [Serial] [Disruptive][NodeFeature:Eviction] when we run containers that should cause DiskPressure should eventually evict all of the correct pods 30s

go run hack/e2e.go -v --test --test_args='--ginkgo.focus=E2eNode\sSuite\s\[k8s\.io\]\sPriorityLocalStorageEvictionOrdering\s\[Slow\]\s\[Serial\]\s\[Disruptive\]\[NodeFeature\:Eviction\]\swhen\swe\srun\scontainers\sthat\sshould\scause\sDiskPressure\s\sshould\seventually\sevict\sall\sof\sthe\scorrect\spods$'
/go/src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:174
Unexpected error:
    <*errors.errorString | 0xc0002329b0>: {
        s: "timed out waiting for the condition",
    }
    timed out waiting for the condition
occurred
/go/src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:236
				
				Click to see stdout/stderrfrom junit_cos-stable2_01.xml

Filter through log files | View test history on testgrid


E2eNode Suite [k8s.io] PriorityMemoryEvictionOrdering [Slow] [Serial] [Disruptive][NodeFeature:Eviction] when we run containers that should cause MemoryPressure should eventually evict all of the correct pods 30s

go run hack/e2e.go -v --test --test_args='--ginkgo.focus=E2eNode\sSuite\s\[k8s\.io\]\sPriorityMemoryEvictionOrdering\s\[Slow\]\s\[Serial\]\s\[Disruptive\]\[NodeFeature\:Eviction\]\swhen\swe\srun\scontainers\sthat\sshould\scause\sMemoryPressure\s\sshould\seventually\sevict\sall\sof\sthe\scorrect\spods$'
/go/src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:174
Unexpected error:
    <*errors.errorString | 0xc0002329b0>: {
        s: "timed out waiting for the condition",
    }
    timed out waiting for the condition
occurred
/go/src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:236
				
				Click to see stdout/stderrfrom junit_cos-stable2_01.xml

Filter through log files | View test history on testgrid


E2eNode Suite [k8s.io] Restart [Serial] [Slow] [Disruptive] [NodeFeature:ContainerRuntimeRestart] Container Runtime Network should recover from ip leak 5m2s

go run hack/e2e.go -v --test --test_args='--ginkgo.focus=E2eNode\sSuite\s\[k8s\.io\]\sRestart\s\[Serial\]\s\[Slow\]\s\[Disruptive\]\s\[NodeFeature\:ContainerRuntimeRestart\]\sContainer\sRuntime\sNetwork\sshould\srecover\sfrom\sip\sleak$'
_output/local/go/src/k8s.io/kubernetes/test/e2e_node/restart_test.go:82
Unexpected error:
    <*url.Error | 0xc002a724b0>: {
        Op: "Get",
        URL: "https://127.0.0.1:6443/api/v1/namespaces/restart-test-6896/pods",
        Err: {
            Op: "dial",
            Net: "tcp",
            Source: nil,
            Addr: {
                IP: [0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 255, 255, 127, 0, 0, 1],
                Port: 6443,
                Zone: "",
            },
            Err: {Syscall: "connect", Err: 0x6f},
        },
    }
    Get "https://127.0.0.1:6443/api/v1/namespaces/restart-test-6896/pods": dial tcp 127.0.0.1:6443: connect: connection refused
occurred
_output/local/go/src/k8s.io/kubernetes/test/e2e_node/resource_collector.go:383
				
				Click to see stdout/stderrfrom junit_cos-stable2_01.xml

Filter through log files | View test history on testgrid


E2eNode Suite [k8s.io] SystemNodeCriticalPod [Slow] [Serial] [Disruptive] [NodeFeature:SystemNodeCriticalPod] when create a system-node-critical pod should not be evicted upon DiskPressure 5m0s

go run hack/e2e.go -v --test --test_args='--ginkgo.focus=E2eNode\sSuite\s\[k8s\.io\]\sSystemNodeCriticalPod\s\[Slow\]\s\[Serial\]\s\[Disruptive\]\s\[NodeFeature\:SystemNodeCriticalPod\]\swhen\screate\sa\ssystem\-node\-critical\spod\s\sshould\snot\sbe\sevicted\supon\sDiskPressure$'
_output/local/go/src/k8s.io/kubernetes/test/e2e_node/util.go:136
Failed to get successful response from /configz
Unexpected error:
    <*errors.errorString | 0xc0002329b0>: {
        s: "timed out waiting for the condition",
    }
    timed out waiting for the condition
occurred
/go/src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/kubelet/config.go:105
				
				Click to see stdout/stderrfrom junit_cos-stable2_01.xml

Filter through log files | View test history on testgrid


E2eNode Suite [k8s.io] [Feature:DynamicKubeletConfig][NodeFeature:DynamicKubeletConfig][Serial][Disruptive] delete and recreate ConfigMap: state transitions: status and events should match expectations 30s

go run hack/e2e.go -v --test --test_args='--ginkgo.focus=E2eNode\sSuite\s\[k8s\.io\]\s\[Feature\:DynamicKubeletConfig\]\[NodeFeature\:DynamicKubeletConfig\]\[Serial\]\[Disruptive\]\s\sdelete\sand\srecreate\sConfigMap\:\sstate\stransitions\:\sstatus\sand\sevents\sshould\smatch\sexpectations$'
/go/src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:174
Unexpected error:
    <*errors.errorString | 0xc0002329b0>: {
        s: "timed out waiting for the condition",
    }
    timed out waiting for the condition
occurred
/go/src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:236
				
				Click to see stdout/stderrfrom junit_cos-stable2_01.xml

Filter through log files | View test history on testgrid


E2eNode Suite [k8s.io] [Feature:DynamicKubeletConfig][NodeFeature:DynamicKubeletConfig][Serial][Disruptive] update ConfigMap in-place: recover to last-known-good version: status and events should match expectations 30s

go run hack/e2e.go -v --test --test_args='--ginkgo.focus=E2eNode\sSuite\s\[k8s\.io\]\s\[Feature\:DynamicKubeletConfig\]\[NodeFeature\:DynamicKubeletConfig\]\[Serial\]\[Disruptive\]\s\supdate\sConfigMap\sin\-place\:\srecover\sto\slast\-known\-good\sversion\:\sstatus\sand\sevents\sshould\smatch\sexpectations$'
/go/src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:174
Unexpected error:
    <*errors.errorString | 0xc0002329b0>: {
        s: "timed out waiting for the condition",
    }
    timed out waiting for the condition
occurred
/go/src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:236
				
				Click to see stdout/stderrfrom junit_cos-stable2_01.xml

Filter through log files | View test history on testgrid


E2eNode Suite [k8s.io] [Feature:DynamicKubeletConfig][NodeFeature:DynamicKubeletConfig][Serial][Disruptive] update ConfigMap in-place: state transitions: status and events should match expectations 30s

go run hack/e2e.go -v --test --test_args='--ginkgo.focus=E2eNode\sSuite\s\[k8s\.io\]\s\[Feature\:DynamicKubeletConfig\]\[NodeFeature\:DynamicKubeletConfig\]\[Serial\]\[Disruptive\]\s\supdate\sConfigMap\sin\-place\:\sstate\stransitions\:\sstatus\sand\sevents\sshould\smatch\sexpectations$'
/go/src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:174
Unexpected error:
    <*errors.errorString | 0xc0002329b0>: {
        s: "timed out waiting for the condition",
    }
    timed out waiting for the condition
occurred
/go/src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:236
				
				Click to see stdout/stderrfrom junit_cos-stable2_01.xml

Filter through log files | View test history on testgrid


E2eNode Suite [k8s.io] [Feature:DynamicKubeletConfig][NodeFeature:DynamicKubeletConfig][Serial][Disruptive] update Node.Spec.ConfigSource: 100 update stress test: status and events should match expectations 30s

go run hack/e2e.go -v --test --test_args='--ginkgo.focus=E2eNode\sSuite\s\[k8s\.io\]\s\[Feature\:DynamicKubeletConfig\]\[NodeFeature\:DynamicKubeletConfig\]\[Serial\]\[Disruptive\]\s\supdate\sNode\.Spec\.ConfigSource\:\s100\supdate\sstress\stest\:\sstatus\sand\sevents\sshould\smatch\sexpectations$'
/go/src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:174
Unexpected error:
    <*errors.errorString | 0xc0002329b0>: {
        s: "timed out waiting for the condition",
    }
    timed out waiting for the condition
occurred
/go/src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:236
				
				Click to see stdout/stderrfrom junit_cos-stable2_01.xml

Filter through log files | View test history on testgrid


E2eNode Suite [k8s.io] [Feature:DynamicKubeletConfig][NodeFeature:DynamicKubeletConfig][Serial][Disruptive] update Node.Spec.ConfigSource: non-nil last-known-good to a new non-nil last-known-good status and events should match expectations 30s

go run hack/e2e.go -v --test --test_args='--ginkgo.focus=E2eNode\sSuite\s\[k8s\.io\]\s\[Feature\:DynamicKubeletConfig\]\[NodeFeature\:DynamicKubeletConfig\]\[Serial\]\[Disruptive\]\s\supdate\sNode\.Spec\.ConfigSource\:\snon\-nil\slast\-known\-good\sto\sa\snew\snon\-nil\slast\-known\-good\sstatus\sand\sevents\sshould\smatch\sexpectations$'
/go/src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:174
Unexpected error:
    <*errors.errorString | 0xc0002329b0>: {
        s: "timed out waiting for the condition",
    }
    timed out waiting for the condition
occurred
/go/src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:236
				
				Click to see stdout/stderrfrom junit_cos-stable2_01.xml

Filter through log files | View test history on testgrid


E2eNode Suite [k8s.io] [Feature:DynamicKubeletConfig][NodeFeature:DynamicKubeletConfig][Serial][Disruptive] update Node.Spec.ConfigSource: recover to last-known-good ConfigMap: status and events should match expectations 30s

go run hack/e2e.go -v --test --test_args='--ginkgo.focus=E2eNode\sSuite\s\[k8s\.io\]\s\[Feature\:DynamicKubeletConfig\]\[NodeFeature\:DynamicKubeletConfig\]\[Serial\]\[Disruptive\]\s\supdate\sNode\.Spec\.ConfigSource\:\srecover\sto\slast\-known\-good\sConfigMap\:\sstatus\sand\sevents\sshould\smatch\sexpectations$'
/go/src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:174
Unexpected error:
    <*errors.errorString | 0xc0002329b0>: {
        s: "timed out waiting for the condition",
    }
    timed out waiting for the condition
occurred
/go/src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:236
				
				Click to see stdout/stderrfrom junit_cos-stable2_01.xml

Filter through log files | View test history on testgrid


E2eNode Suite [k8s.io] [Feature:DynamicKubeletConfig][NodeFeature:DynamicKubeletConfig][Serial][Disruptive] update Node.Spec.ConfigSource: state transitions: status and events should match expectations 30s

go run hack/e2e.go -v --test --test_args='--ginkgo.focus=E2eNode\sSuite\s\[k8s\.io\]\s\[Feature\:DynamicKubeletConfig\]\[NodeFeature\:DynamicKubeletConfig\]\[Serial\]\[Disruptive\]\s\supdate\sNode\.Spec\.ConfigSource\:\sstate\stransitions\:\sstatus\sand\sevents\sshould\smatch\sexpectations$'
/go/src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:174
Unexpected error:
    <*errors.errorString | 0xc0002329b0>: {
        s: "timed out waiting for the condition",
    }
    timed out waiting for the condition
occurred
/go/src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:236
				
				Click to see stdout/stderrfrom junit_cos-stable2_01.xml

Filter through log files | View test history on testgrid


E2eNode Suite [sig-node] Dockershim [Serial] [Disruptive] [Feature:Docker][Legacy:Docker] When all containers in pod are missing should complete pod sandbox clean up based on the information in sandbox checkpoint 30s

go run hack/e2e.go -v --test --test_args='--ginkgo.focus=E2eNode\sSuite\s\[sig\-node\]\sDockershim\s\[Serial\]\s\[Disruptive\]\s\[Feature\:Docker\]\[Legacy\:Docker\]\sWhen\sall\scontainers\sin\spod\sare\smissing\sshould\scomplete\spod\ssandbox\sclean\sup\sbased\son\sthe\sinformation\sin\ssandbox\scheckpoint$'
/go/src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:174
Unexpected error:
    <*errors.errorString | 0xc0002329b0>: {
        s: "timed out waiting for the condition",
    }
    timed out waiting for the condition
occurred
/go/src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:236
				
				Click to see stdout/stderrfrom junit_cos-stable2_01.xml

Filter through log files | View test history on testgrid


E2eNode Suite [sig-node] Dockershim [Serial] [Disruptive] [Feature:Docker][Legacy:Docker] When checkpoint file is corrupted should complete pod sandbox clean up 30s

go run hack/e2e.go -v --test --test_args='--ginkgo.focus=E2eNode\sSuite\s\[sig\-node\]\sDockershim\s\[Serial\]\s\[Disruptive\]\s\[Feature\:Docker\]\[Legacy\:Docker\]\sWhen\scheckpoint\sfile\sis\scorrupted\sshould\scomplete\spod\ssandbox\sclean\sup$'
/go/src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:174
Unexpected error:
    <*errors.errorString | 0xc0002329b0>: {
        s: "timed out waiting for the condition",
    }
    timed out waiting for the condition
occurred
/go/src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:236
				
				Click to see stdout/stderrfrom junit_cos-stable2_01.xml

Filter through log files | View test history on testgrid


E2eNode Suite [sig-node] Dockershim [Serial] [Disruptive] [Feature:Docker][Legacy:Docker] When pod sandbox checkpoint is missing should complete pod sandbox clean up 30s

go run hack/e2e.go -v --test --test_args='--ginkgo.focus=E2eNode\sSuite\s\[sig\-node\]\sDockershim\s\[Serial\]\s\[Disruptive\]\s\[Feature\:Docker\]\[Legacy\:Docker\]\sWhen\spod\ssandbox\scheckpoint\sis\smissing\sshould\scomplete\spod\ssandbox\sclean\sup$'
/go/src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:174
Unexpected error:
    <*errors.errorString | 0xc0002329b0>: {
        s: "timed out waiting for the condition",
    }
    timed out waiting for the condition
occurred
/go/src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:236
				
				Click to see stdout/stderrfrom junit_cos-stable2_01.xml

Filter through log files | View test history on testgrid


E2eNode Suite [sig-node] Dockershim [Serial] [Disruptive] [Feature:Docker][Legacy:Docker] should remove dangling checkpoint file 30s

go run hack/e2e.go -v --test --test_args='--ginkgo.focus=E2eNode\sSuite\s\[sig\-node\]\sDockershim\s\[Serial\]\s\[Disruptive\]\s\[Feature\:Docker\]\[Legacy\:Docker\]\sshould\sremove\sdangling\scheckpoint\sfile$'
/go/src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:174
Unexpected error:
    <*errors.errorString | 0xc0002329b0>: {
        s: "timed out waiting for the condition",
    }
    timed out waiting for the condition
occurred
/go/src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:236
				
				Click to see stdout/stderrfrom junit_cos-stable2_01.xml

Filter through log files | View test history on testgrid


E2eNode Suite [sig-node] PodPidsLimit [Serial] With config updated with pids limits should set pids.max for Pod 30s

go run hack/e2e.go -v --test --test_args='--ginkgo.focus=E2eNode\sSuite\s\[sig\-node\]\sPodPidsLimit\s\[Serial\]\sWith\sconfig\supdated\swith\spids\slimits\sshould\sset\spids\.max\sfor\sPod$'
/go/src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:174
Unexpected error:
    <*errors.errorString | 0xc0002329b0>: {
        s: "timed out waiting for the condition",
    }
    timed out waiting for the condition
occurred
/go/src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:236
				
				Click to see stdout/stderrfrom junit_cos-stable2_01.xml

Filter through log files | View test history on testgrid


E2eNode Suite [sig-node] Resource-usage [Serial] [Slow] regular resource usage tracking resource tracking for 10 pods per node 30s

go run hack/e2e.go -v --test --test_args='--ginkgo.focus=E2eNode\sSuite\s\[sig\-node\]\sResource\-usage\s\[Serial\]\s\[Slow\]\sregular\sresource\susage\stracking\sresource\stracking\sfor\s10\spods\sper\snode$'
/go/src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:174
Unexpected error:
    <*errors.errorString | 0xc0002329b0>: {
        s: "timed out waiting for the condition",
    }
    timed out waiting for the condition
occurred
/go/src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:236
				
				Click to see stdout/stderrfrom junit_cos-stable2_01.xml

Filter through log files | View test history on testgrid


E2eNode Suite [sig-node] Topology Manager [Serial] [Feature:TopologyManager][NodeFeature:TopologyManager] With kubeconfig updated to static CPU Manager policy run the Topology Manager tests run Topology Manager node alignment test suite 30s

go run hack/e2e.go -v --test --test_args='--ginkgo.focus=E2eNode\sSuite\s\[sig\-node\]\sTopology\sManager\s\[Serial\]\s\[Feature\:TopologyManager\]\[NodeFeature\:TopologyManager\]\sWith\skubeconfig\supdated\sto\sstatic\sCPU\sManager\spolicy\srun\sthe\sTopology\sManager\stests\srun\sTopology\sManager\snode\salignment\stest\ssuite$'
/go/src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:174
Unexpected error:
    <*errors.errorString | 0xc0002329b0>: {
        s: "timed out waiting for the condition",
    }
    timed out waiting for the condition
occurred
/go/src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:236
				
				Click to see stdout/stderrfrom junit_cos-stable2_01.xml

Filter through log files | View test history on testgrid


E2eNode Suite [sig-node] Topology Manager [Serial] [Feature:TopologyManager][NodeFeature:TopologyManager] With kubeconfig updated to static CPU Manager policy run the Topology Manager tests run Topology Manager policy test suite 30s

go run hack/e2e.go -v --test --test_args='--ginkgo.focus=E2eNode\sSuite\s\[sig\-node\]\sTopology\sManager\s\[Serial\]\s\[Feature\:TopologyManager\]\[NodeFeature\:TopologyManager\]\sWith\skubeconfig\supdated\sto\sstatic\sCPU\sManager\spolicy\srun\sthe\sTopology\sManager\stests\srun\sTopology\sManager\spolicy\stest\ssuite$'
/go/src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:174
Unexpected error:
    <*errors.errorString | 0xc0002329b0>: {
        s: "timed out waiting for the condition",
    }
    timed out waiting for the condition
occurred
/go/src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:236
				
				Click to see stdout/stderrfrom junit_cos-stable2_01.xml

Filter through log files | View test history on testgrid


Node Tests 4h59m

error during go run /go/src/k8s.io/kubernetes/test/e2e_node/runner/remote/run_remote.go --cleanup --logtostderr --vmodule=*=4 --ssh-env=gce --results-dir=/workspace/_artifacts --project=k8s-jkns-gke-ubuntu-1-6 --zone=us-west1-b --ssh-user=prow --ssh-key=/workspace/.ssh/google_compute_engine --ginkgo-flags=--nodes=1 --focus="\[Serial\]" --skip="\[Flaky\]|\[Benchmark\]|\[NodeSpecialFeature:.+\]|\[NodeAlphaFeature:.+\]" --test_args=--feature-gates=DynamicKubeletConfig=true,LocalStorageCapacityIsolation=true --kubelet-flags="--cgroups-per-qos=true --cgroup-root=/" --test-timeout=5h0m0s --image-config-file=/workspace/test-infra/jobs/e2e_node/image-config-serial.yaml (interrupted): exit status 1
				from junit_runner.xml

Filter through log files | View test history on testgrid


Timeout 5h0m

kubetest --timeout triggered
				from junit_runner.xml

Filter through log files | View test history on testgrid


Show 14 Passed Tests

Show 280 Skipped Tests