This job view page is being replaced by Spyglass soon. Check out the new job view.
PRjohngmyers: Fix certificate bootstrap for non-kops-controller-bootstrap cloud providers
ResultABORTED
Tests 0 failed / 0 succeeded
Started2021-07-18 19:35
Elapsed36m27s
Revision302a5c4d0b79909a4e1326b6114da4b65dac4e47
Refs 12019

No Test Failures!


Error lines from build-log.txt

... skipping 487 lines ...
I0718 19:40:02.360071    4255 copy.go:30] cp /home/prow/go/src/k8s.io/kops/bazel-bin/cmd/kops/linux-amd64/kops /logs/artifacts/4418b825-e7ff-11eb-b28a-da9364982371/kops
I0718 19:40:02.939949    4255 up.go:43] Cleaning up any leaked resources from previous cluster
I0718 19:40:02.940931    4255 dumplogs.go:38] /home/prow/go/src/k8s.io/kops/bazel-bin/cmd/kops/linux-amd64/kops toolbox dump --name e2e-test-do.k8s.local --dir /logs/artifacts --private-key /etc/do-ssh/private-ssh-key --ssh-user root
I0718 19:40:02.961256   11746 featureflag.go:167] FeatureFlag "SpecOverrideFlag"=true
I0718 19:40:02.961447   11746 featureflag.go:167] FeatureFlag "AlphaAllowGCE"=true
I0718 19:40:03.009992   11746 s3context.go:92] Found S3_ENDPOINT="sfo3.digitaloceanspaces.com", using as non-AWS S3 backend
Error: Cluster.kops.k8s.io "e2e-test-do.k8s.local" not found

Cluster.kops.k8s.io "e2e-test-do.k8s.local" not found
W0718 19:40:03.201617    4255 down.go:34] Dumping cluster logs at the start of Down() failed: exit status 1
I0718 19:40:03.202821    4255 down.go:48] /home/prow/go/src/k8s.io/kops/bazel-bin/cmd/kops/linux-amd64/kops delete cluster --name e2e-test-do.k8s.local --yes
I0718 19:40:03.219860   11756 featureflag.go:167] FeatureFlag "SpecOverrideFlag"=true
I0718 19:40:03.220089   11756 featureflag.go:167] FeatureFlag "AlphaAllowGCE"=true
I0718 19:40:03.265929   11756 s3context.go:92] Found S3_ENDPOINT="sfo3.digitaloceanspaces.com", using as non-AWS S3 backend
Error: error reading cluster configuration: Cluster.kops.k8s.io "e2e-test-do.k8s.local" not found

error reading cluster configuration: Cluster.kops.k8s.io "e2e-test-do.k8s.local" not found
I0718 19:40:03.460799    4255 http.go:37] curl http://metadata.google.internal/computeMetadata/v1/instance/network-interfaces/0/access-configs/0/external-ip
2021/07/18 19:40:03 failed to get external ip from metadata service: http://metadata.google.internal/computeMetadata/v1/instance/network-interfaces/0/access-configs/0/external-ip returned 404
I0718 19:40:03.480725    4255 http.go:37] curl https://ip.jsb.workers.dev
I0718 19:40:03.590089    4255 up.go:144] /home/prow/go/src/k8s.io/kops/bazel-bin/cmd/kops/linux-amd64/kops create cluster --name e2e-test-do.k8s.local --cloud digitalocean --kubernetes-version https://storage.googleapis.com/kubernetes-release/release/v1.21.3 --ssh-public-key /etc/do-ssh/public-ssh-key --override cluster.spec.nodePortAccess=0.0.0.0/0 --yes --networking=calico --api-loadbalancer-type=public --node-count=2 --master-count=3 --admin-access 35.193.210.107/32 --master-volume-size 48 --node-volume-size 48 --zones nyc1 --master-size c2-16vcpu-32gb --node-size c2-16vcpu-32gb
I0718 19:40:03.607319   11766 featureflag.go:167] FeatureFlag "SpecOverrideFlag"=true
I0718 19:40:03.607406   11766 featureflag.go:167] FeatureFlag "AlphaAllowGCE"=true
I0718 19:40:03.647290   11766 create_cluster.go:825] Using SSH public key: /etc/do-ssh/public-ssh-key
I0718 19:40:03.647391   11766 s3context.go:92] Found S3_ENDPOINT="sfo3.digitaloceanspaces.com", using as non-AWS S3 backend
... skipping 25 lines ...
I0718 19:40:06.933177   11766 executor.go:111] Tasks: 41 done / 57 total; 7 can run
I0718 19:40:07.021255   11766 keypair.go:213] Issuing new certificate: "kubelet"
I0718 19:40:07.022492   11766 keypair.go:213] Issuing new certificate: "kube-proxy"
W0718 19:40:07.088387   11766 loadbalancer.go:217] IP address for LB api-e2e-test-do-k8s-local not yet available -- sleeping 10s
W0718 19:40:07.175724   11766 loadbalancer.go:217] IP address for LB api-e2e-test-do-k8s-local not yet available -- sleeping 10s
W0718 19:40:07.679502   11766 loadbalancer.go:217] IP address for LB api-e2e-test-do-k8s-local not yet available -- sleeping 10s
W0718 19:40:17.679820   11766 executor.go:139] error running task "BootstrapScript/master-nyc1-1" (9m49s remaining to succeed): error finding address for *dotasks.LoadBalancer {"Name":"api-e2e-test-do-k8s-local","ID":"d0fa1b62-29b3-4533-ad6b-5dd13d884d96","Lifecycle":"Sync","Region":"nyc1","DropletTag":"KubernetesCluster-Master:e2e-test-do-k8s-local","IPAddress":"","ForAPIServer":true}: IP Address is still empty.
W0718 19:40:17.679873   11766 executor.go:139] error running task "BootstrapScript/master-nyc1-3" (9m49s remaining to succeed): error finding address for *dotasks.LoadBalancer {"Name":"api-e2e-test-do-k8s-local","ID":"d0fa1b62-29b3-4533-ad6b-5dd13d884d96","Lifecycle":"Sync","Region":"nyc1","DropletTag":"KubernetesCluster-Master:e2e-test-do-k8s-local","IPAddress":"","ForAPIServer":true}: IP Address is still empty.
W0718 19:40:17.679893   11766 executor.go:139] error running task "BootstrapScript/master-nyc1-2" (9m49s remaining to succeed): error finding address for *dotasks.LoadBalancer {"Name":"api-e2e-test-do-k8s-local","ID":"d0fa1b62-29b3-4533-ad6b-5dd13d884d96","Lifecycle":"Sync","Region":"nyc1","DropletTag":"KubernetesCluster-Master:e2e-test-do-k8s-local","IPAddress":"","ForAPIServer":true}: IP Address is still empty.
I0718 19:40:17.679913   11766 executor.go:111] Tasks: 45 done / 57 total; 4 can run
W0718 19:40:17.909384   11766 loadbalancer.go:217] IP address for LB api-e2e-test-do-k8s-local not yet available -- sleeping 10s
W0718 19:40:17.942784   11766 loadbalancer.go:217] IP address for LB api-e2e-test-do-k8s-local not yet available -- sleeping 10s
W0718 19:40:18.024162   11766 loadbalancer.go:217] IP address for LB api-e2e-test-do-k8s-local not yet available -- sleeping 10s
W0718 19:40:18.071489   11766 loadbalancer.go:217] IP address for LB api-e2e-test-do-k8s-local not yet available -- sleeping 10s
W0718 19:40:28.072476   11766 executor.go:139] error running task "BootstrapScript/nodes-nyc1" (9m49s remaining to succeed): error finding address for *dotasks.LoadBalancer {"Name":"api-e2e-test-do-k8s-local","ID":"d0fa1b62-29b3-4533-ad6b-5dd13d884d96","Lifecycle":"Sync","Region":"nyc1","DropletTag":"KubernetesCluster-Master:e2e-test-do-k8s-local","IPAddress":"","ForAPIServer":true}: IP Address is still empty.
W0718 19:40:28.072522   11766 executor.go:139] error running task "BootstrapScript/master-nyc1-1" (9m38s remaining to succeed): error finding address for *dotasks.LoadBalancer {"Name":"api-e2e-test-do-k8s-local","ID":"d0fa1b62-29b3-4533-ad6b-5dd13d884d96","Lifecycle":"Sync","Region":"nyc1","DropletTag":"KubernetesCluster-Master:e2e-test-do-k8s-local","IPAddress":"","ForAPIServer":true}: IP Address is still empty.
W0718 19:40:28.072534   11766 executor.go:139] error running task "BootstrapScript/master-nyc1-3" (9m38s remaining to succeed): error finding address for *dotasks.LoadBalancer {"Name":"api-e2e-test-do-k8s-local","ID":"d0fa1b62-29b3-4533-ad6b-5dd13d884d96","Lifecycle":"Sync","Region":"nyc1","DropletTag":"KubernetesCluster-Master:e2e-test-do-k8s-local","IPAddress":"","ForAPIServer":true}: IP Address is still empty.
W0718 19:40:28.072542   11766 executor.go:139] error running task "BootstrapScript/master-nyc1-2" (9m38s remaining to succeed): error finding address for *dotasks.LoadBalancer {"Name":"api-e2e-test-do-k8s-local","ID":"d0fa1b62-29b3-4533-ad6b-5dd13d884d96","Lifecycle":"Sync","Region":"nyc1","DropletTag":"KubernetesCluster-Master:e2e-test-do-k8s-local","IPAddress":"","ForAPIServer":true}: IP Address is still empty.
I0718 19:40:28.072548   11766 executor.go:155] No progress made, sleeping before retrying 4 task(s)
I0718 19:40:38.072833   11766 executor.go:111] Tasks: 45 done / 57 total; 4 can run
E0718 19:40:38.087148   11766 loadbalancer.go:185] Error fetching load balancer with Name=api-e2e-test-do-k8s-local
E0718 19:40:38.087236   11766 loadbalancer.go:185] Error fetching load balancer with Name=api-e2e-test-do-k8s-local
E0718 19:40:38.087255   11766 loadbalancer.go:185] Error fetching load balancer with Name=api-e2e-test-do-k8s-local
E0718 19:40:38.087275   11766 loadbalancer.go:185] Error fetching load balancer with Name=api-e2e-test-do-k8s-local
W0718 19:40:38.087361   11766 executor.go:139] error running task "BootstrapScript/nodes-nyc1" (9m39s remaining to succeed): error finding address for *dotasks.LoadBalancer {"Name":"api-e2e-test-do-k8s-local","ID":"d0fa1b62-29b3-4533-ad6b-5dd13d884d96","Lifecycle":"Sync","Region":"nyc1","DropletTag":"KubernetesCluster-Master:e2e-test-do-k8s-local","IPAddress":"","ForAPIServer":true}: Get "https://api.digitalocean.com/v2/load_balancers/d0fa1b62-29b3-4533-ad6b-5dd13d884d96": read tcp 10.60.100.198:44698->104.16.182.15:443: read: connection reset by peer
W0718 19:40:38.087400   11766 executor.go:139] error running task "BootstrapScript/master-nyc1-1" (9m28s remaining to succeed): error finding address for *dotasks.LoadBalancer {"Name":"api-e2e-test-do-k8s-local","ID":"d0fa1b62-29b3-4533-ad6b-5dd13d884d96","Lifecycle":"Sync","Region":"nyc1","DropletTag":"KubernetesCluster-Master:e2e-test-do-k8s-local","IPAddress":"","ForAPIServer":true}: Get "https://api.digitalocean.com/v2/load_balancers/d0fa1b62-29b3-4533-ad6b-5dd13d884d96": read tcp 10.60.100.198:44698->104.16.182.15:443: read: connection reset by peer
W0718 19:40:38.087418   11766 executor.go:139] error running task "BootstrapScript/master-nyc1-3" (9m28s remaining to succeed): error finding address for *dotasks.LoadBalancer {"Name":"api-e2e-test-do-k8s-local","ID":"d0fa1b62-29b3-4533-ad6b-5dd13d884d96","Lifecycle":"Sync","Region":"nyc1","DropletTag":"KubernetesCluster-Master:e2e-test-do-k8s-local","IPAddress":"","ForAPIServer":true}: Get "https://api.digitalocean.com/v2/load_balancers/d0fa1b62-29b3-4533-ad6b-5dd13d884d96": read tcp 10.60.100.198:44698->104.16.182.15:443: read: connection reset by peer
W0718 19:40:38.087431   11766 executor.go:139] error running task "BootstrapScript/master-nyc1-2" (9m28s remaining to succeed): error finding address for *dotasks.LoadBalancer {"Name":"api-e2e-test-do-k8s-local","ID":"d0fa1b62-29b3-4533-ad6b-5dd13d884d96","Lifecycle":"Sync","Region":"nyc1","DropletTag":"KubernetesCluster-Master:e2e-test-do-k8s-local","IPAddress":"","ForAPIServer":true}: Get "https://api.digitalocean.com/v2/load_balancers/d0fa1b62-29b3-4533-ad6b-5dd13d884d96": read tcp 10.60.100.198:44698->104.16.182.15:443: read: connection reset by peer
I0718 19:40:38.087440   11766 executor.go:155] No progress made, sleeping before retrying 4 task(s)
I0718 19:40:48.091652   11766 executor.go:111] Tasks: 45 done / 57 total; 4 can run
W0718 19:40:48.335492   11766 loadbalancer.go:217] IP address for LB api-e2e-test-do-k8s-local not yet available -- sleeping 10s
W0718 19:40:48.363669   11766 loadbalancer.go:217] IP address for LB api-e2e-test-do-k8s-local not yet available -- sleeping 10s
W0718 19:40:48.389580   11766 loadbalancer.go:217] IP address for LB api-e2e-test-do-k8s-local not yet available -- sleeping 10s
W0718 19:40:48.467668   11766 loadbalancer.go:217] IP address for LB api-e2e-test-do-k8s-local not yet available -- sleeping 10s
W0718 19:40:58.469029   11766 executor.go:139] error running task "BootstrapScript/master-nyc1-2" (9m8s remaining to succeed): error finding address for *dotasks.LoadBalancer {"Name":"api-e2e-test-do-k8s-local","ID":"d0fa1b62-29b3-4533-ad6b-5dd13d884d96","Lifecycle":"Sync","Region":"nyc1","DropletTag":"KubernetesCluster-Master:e2e-test-do-k8s-local","IPAddress":"","ForAPIServer":true}: IP Address is still empty.
W0718 19:40:58.469090   11766 executor.go:139] error running task "BootstrapScript/nodes-nyc1" (9m19s remaining to succeed): error finding address for *dotasks.LoadBalancer {"Name":"api-e2e-test-do-k8s-local","ID":"d0fa1b62-29b3-4533-ad6b-5dd13d884d96","Lifecycle":"Sync","Region":"nyc1","DropletTag":"KubernetesCluster-Master:e2e-test-do-k8s-local","IPAddress":"","ForAPIServer":true}: IP Address is still empty.
W0718 19:40:58.469104   11766 executor.go:139] error running task "BootstrapScript/master-nyc1-1" (9m8s remaining to succeed): error finding address for *dotasks.LoadBalancer {"Name":"api-e2e-test-do-k8s-local","ID":"d0fa1b62-29b3-4533-ad6b-5dd13d884d96","Lifecycle":"Sync","Region":"nyc1","DropletTag":"KubernetesCluster-Master:e2e-test-do-k8s-local","IPAddress":"","ForAPIServer":true}: IP Address is still empty.
W0718 19:40:58.469113   11766 executor.go:139] error running task "BootstrapScript/master-nyc1-3" (9m8s remaining to succeed): error finding address for *dotasks.LoadBalancer {"Name":"api-e2e-test-do-k8s-local","ID":"d0fa1b62-29b3-4533-ad6b-5dd13d884d96","Lifecycle":"Sync","Region":"nyc1","DropletTag":"KubernetesCluster-Master:e2e-test-do-k8s-local","IPAddress":"","ForAPIServer":true}: IP Address is still empty.
I0718 19:40:58.469120   11766 executor.go:155] No progress made, sleeping before retrying 4 task(s)
I0718 19:41:08.470078   11766 executor.go:111] Tasks: 45 done / 57 total; 4 can run
W0718 19:41:08.703369   11766 loadbalancer.go:217] IP address for LB api-e2e-test-do-k8s-local not yet available -- sleeping 10s
W0718 19:41:08.861886   11766 loadbalancer.go:217] IP address for LB api-e2e-test-do-k8s-local not yet available -- sleeping 10s
W0718 19:41:08.872527   11766 loadbalancer.go:217] IP address for LB api-e2e-test-do-k8s-local not yet available -- sleeping 10s
W0718 19:41:08.944184   11766 loadbalancer.go:217] IP address for LB api-e2e-test-do-k8s-local not yet available -- sleeping 10s
W0718 19:41:18.945093   11766 executor.go:139] error running task "BootstrapScript/master-nyc1-2" (8m47s remaining to succeed): error finding address for *dotasks.LoadBalancer {"Name":"api-e2e-test-do-k8s-local","ID":"d0fa1b62-29b3-4533-ad6b-5dd13d884d96","Lifecycle":"Sync","Region":"nyc1","DropletTag":"KubernetesCluster-Master:e2e-test-do-k8s-local","IPAddress":"","ForAPIServer":true}: IP Address is still empty.
W0718 19:41:18.945140   11766 executor.go:139] error running task "BootstrapScript/nodes-nyc1" (8m58s remaining to succeed): error finding address for *dotasks.LoadBalancer {"Name":"api-e2e-test-do-k8s-local","ID":"d0fa1b62-29b3-4533-ad6b-5dd13d884d96","Lifecycle":"Sync","Region":"nyc1","DropletTag":"KubernetesCluster-Master:e2e-test-do-k8s-local","IPAddress":"","ForAPIServer":true}: IP Address is still empty.
W0718 19:41:18.945155   11766 executor.go:139] error running task "BootstrapScript/master-nyc1-1" (8m47s remaining to succeed): error finding address for *dotasks.LoadBalancer {"Name":"api-e2e-test-do-k8s-local","ID":"d0fa1b62-29b3-4533-ad6b-5dd13d884d96","Lifecycle":"Sync","Region":"nyc1","DropletTag":"KubernetesCluster-Master:e2e-test-do-k8s-local","IPAddress":"","ForAPIServer":true}: IP Address is still empty.
W0718 19:41:18.945165   11766 executor.go:139] error running task "BootstrapScript/master-nyc1-3" (8m47s remaining to succeed): error finding address for *dotasks.LoadBalancer {"Name":"api-e2e-test-do-k8s-local","ID":"d0fa1b62-29b3-4533-ad6b-5dd13d884d96","Lifecycle":"Sync","Region":"nyc1","DropletTag":"KubernetesCluster-Master:e2e-test-do-k8s-local","IPAddress":"","ForAPIServer":true}: IP Address is still empty.
I0718 19:41:18.945172   11766 executor.go:155] No progress made, sleeping before retrying 4 task(s)
I0718 19:41:28.945410   11766 executor.go:111] Tasks: 45 done / 57 total; 4 can run
W0718 19:41:29.245222   11766 loadbalancer.go:217] IP address for LB api-e2e-test-do-k8s-local not yet available -- sleeping 10s
W0718 19:41:29.294232   11766 loadbalancer.go:217] IP address for LB api-e2e-test-do-k8s-local not yet available -- sleeping 10s
W0718 19:41:29.304097   11766 loadbalancer.go:217] IP address for LB api-e2e-test-do-k8s-local not yet available -- sleeping 10s
W0718 19:41:29.340221   11766 loadbalancer.go:217] IP address for LB api-e2e-test-do-k8s-local not yet available -- sleeping 10s
W0718 19:41:39.341147   11766 executor.go:139] error running task "BootstrapScript/nodes-nyc1" (8m38s remaining to succeed): error finding address for *dotasks.LoadBalancer {"Name":"api-e2e-test-do-k8s-local","ID":"d0fa1b62-29b3-4533-ad6b-5dd13d884d96","Lifecycle":"Sync","Region":"nyc1","DropletTag":"KubernetesCluster-Master:e2e-test-do-k8s-local","IPAddress":"","ForAPIServer":true}: IP Address is still empty.
W0718 19:41:39.341196   11766 executor.go:139] error running task "BootstrapScript/master-nyc1-1" (8m27s remaining to succeed): error finding address for *dotasks.LoadBalancer {"Name":"api-e2e-test-do-k8s-local","ID":"d0fa1b62-29b3-4533-ad6b-5dd13d884d96","Lifecycle":"Sync","Region":"nyc1","DropletTag":"KubernetesCluster-Master:e2e-test-do-k8s-local","IPAddress":"","ForAPIServer":true}: IP Address is still empty.
W0718 19:41:39.341209   11766 executor.go:139] error running task "BootstrapScript/master-nyc1-3" (8m27s remaining to succeed): error finding address for *dotasks.LoadBalancer {"Name":"api-e2e-test-do-k8s-local","ID":"d0fa1b62-29b3-4533-ad6b-5dd13d884d96","Lifecycle":"Sync","Region":"nyc1","DropletTag":"KubernetesCluster-Master:e2e-test-do-k8s-local","IPAddress":"","ForAPIServer":true}: IP Address is still empty.
W0718 19:41:39.341217   11766 executor.go:139] error running task "BootstrapScript/master-nyc1-2" (8m27s remaining to succeed): error finding address for *dotasks.LoadBalancer {"Name":"api-e2e-test-do-k8s-local","ID":"d0fa1b62-29b3-4533-ad6b-5dd13d884d96","Lifecycle":"Sync","Region":"nyc1","DropletTag":"KubernetesCluster-Master:e2e-test-do-k8s-local","IPAddress":"","ForAPIServer":true}: IP Address is still empty.
I0718 19:41:39.341257   11766 executor.go:155] No progress made, sleeping before retrying 4 task(s)
I0718 19:41:49.341434   11766 executor.go:111] Tasks: 45 done / 57 total; 4 can run
W0718 19:41:49.485386   11766 loadbalancer.go:217] IP address for LB api-e2e-test-do-k8s-local not yet available -- sleeping 10s
W0718 19:41:49.579430   11766 loadbalancer.go:217] IP address for LB api-e2e-test-do-k8s-local not yet available -- sleeping 10s
W0718 19:41:49.957210   11766 loadbalancer.go:217] IP address for LB api-e2e-test-do-k8s-local not yet available -- sleeping 10s
W0718 19:41:50.110570   11766 loadbalancer.go:217] IP address for LB api-e2e-test-do-k8s-local not yet available -- sleeping 10s
W0718 19:42:00.111455   11766 executor.go:139] error running task "BootstrapScript/master-nyc1-1" (8m6s remaining to succeed): error finding address for *dotasks.LoadBalancer {"Name":"api-e2e-test-do-k8s-local","ID":"d0fa1b62-29b3-4533-ad6b-5dd13d884d96","Lifecycle":"Sync","Region":"nyc1","DropletTag":"KubernetesCluster-Master:e2e-test-do-k8s-local","IPAddress":"","ForAPIServer":true}: IP Address is still empty.
W0718 19:42:00.111509   11766 executor.go:139] error running task "BootstrapScript/master-nyc1-3" (8m6s remaining to succeed): error finding address for *dotasks.LoadBalancer {"Name":"api-e2e-test-do-k8s-local","ID":"d0fa1b62-29b3-4533-ad6b-5dd13d884d96","Lifecycle":"Sync","Region":"nyc1","DropletTag":"KubernetesCluster-Master:e2e-test-do-k8s-local","IPAddress":"","ForAPIServer":true}: IP Address is still empty.
W0718 19:42:00.111526   11766 executor.go:139] error running task "BootstrapScript/master-nyc1-2" (8m6s remaining to succeed): error finding address for *dotasks.LoadBalancer {"Name":"api-e2e-test-do-k8s-local","ID":"d0fa1b62-29b3-4533-ad6b-5dd13d884d96","Lifecycle":"Sync","Region":"nyc1","DropletTag":"KubernetesCluster-Master:e2e-test-do-k8s-local","IPAddress":"","ForAPIServer":true}: IP Address is still empty.
W0718 19:42:00.111535   11766 executor.go:139] error running task "BootstrapScript/nodes-nyc1" (8m17s remaining to succeed): error finding address for *dotasks.LoadBalancer {"Name":"api-e2e-test-do-k8s-local","ID":"d0fa1b62-29b3-4533-ad6b-5dd13d884d96","Lifecycle":"Sync","Region":"nyc1","DropletTag":"KubernetesCluster-Master:e2e-test-do-k8s-local","IPAddress":"","ForAPIServer":true}: IP Address is still empty.
I0718 19:42:00.111542   11766 executor.go:155] No progress made, sleeping before retrying 4 task(s)
I0718 19:42:10.111842   11766 executor.go:111] Tasks: 45 done / 57 total; 4 can run
W0718 19:42:10.348810   11766 loadbalancer.go:217] IP address for LB api-e2e-test-do-k8s-local not yet available -- sleeping 10s
W0718 19:42:10.369724   11766 loadbalancer.go:217] IP address for LB api-e2e-test-do-k8s-local not yet available -- sleeping 10s
W0718 19:42:10.391943   11766 loadbalancer.go:217] IP address for LB api-e2e-test-do-k8s-local not yet available -- sleeping 10s
W0718 19:42:10.424916   11766 loadbalancer.go:217] IP address for LB api-e2e-test-do-k8s-local not yet available -- sleeping 10s
W0718 19:42:20.426079   11766 executor.go:139] error running task "BootstrapScript/nodes-nyc1" (7m57s remaining to succeed): error finding address for *dotasks.LoadBalancer {"Name":"api-e2e-test-do-k8s-local","ID":"d0fa1b62-29b3-4533-ad6b-5dd13d884d96","Lifecycle":"Sync","Region":"nyc1","DropletTag":"KubernetesCluster-Master:e2e-test-do-k8s-local","IPAddress":"","ForAPIServer":true}: IP Address is still empty.
W0718 19:42:20.426129   11766 executor.go:139] error running task "BootstrapScript/master-nyc1-1" (7m46s remaining to succeed): error finding address for *dotasks.LoadBalancer {"Name":"api-e2e-test-do-k8s-local","ID":"d0fa1b62-29b3-4533-ad6b-5dd13d884d96","Lifecycle":"Sync","Region":"nyc1","DropletTag":"KubernetesCluster-Master:e2e-test-do-k8s-local","IPAddress":"","ForAPIServer":true}: IP Address is still empty.
W0718 19:42:20.426145   11766 executor.go:139] error running task "BootstrapScript/master-nyc1-3" (7m46s remaining to succeed): error finding address for *dotasks.LoadBalancer {"Name":"api-e2e-test-do-k8s-local","ID":"d0fa1b62-29b3-4533-ad6b-5dd13d884d96","Lifecycle":"Sync","Region":"nyc1","DropletTag":"KubernetesCluster-Master:e2e-test-do-k8s-local","IPAddress":"","ForAPIServer":true}: IP Address is still empty.
W0718 19:42:20.426154   11766 executor.go:139] error running task "BootstrapScript/master-nyc1-2" (7m46s remaining to succeed): error finding address for *dotasks.LoadBalancer {"Name":"api-e2e-test-do-k8s-local","ID":"d0fa1b62-29b3-4533-ad6b-5dd13d884d96","Lifecycle":"Sync","Region":"nyc1","DropletTag":"KubernetesCluster-Master:e2e-test-do-k8s-local","IPAddress":"","ForAPIServer":true}: IP Address is still empty.
I0718 19:42:20.426161   11766 executor.go:155] No progress made, sleeping before retrying 4 task(s)
I0718 19:42:30.427817   11766 executor.go:111] Tasks: 45 done / 57 total; 4 can run
W0718 19:42:30.637911   11766 loadbalancer.go:217] IP address for LB api-e2e-test-do-k8s-local not yet available -- sleeping 10s
W0718 19:42:30.710932   11766 loadbalancer.go:217] IP address for LB api-e2e-test-do-k8s-local not yet available -- sleeping 10s
W0718 19:42:30.774429   11766 loadbalancer.go:217] IP address for LB api-e2e-test-do-k8s-local not yet available -- sleeping 10s
W0718 19:42:30.833881   11766 loadbalancer.go:217] IP address for LB api-e2e-test-do-k8s-local not yet available -- sleeping 10s
W0718 19:42:40.834065   11766 executor.go:139] error running task "BootstrapScript/master-nyc1-1" (7m26s remaining to succeed): error finding address for *dotasks.LoadBalancer {"Name":"api-e2e-test-do-k8s-local","ID":"d0fa1b62-29b3-4533-ad6b-5dd13d884d96","Lifecycle":"Sync","Region":"nyc1","DropletTag":"KubernetesCluster-Master:e2e-test-do-k8s-local","IPAddress":"","ForAPIServer":true}: IP Address is still empty.
W0718 19:42:40.834140   11766 executor.go:139] error running task "BootstrapScript/master-nyc1-3" (7m26s remaining to succeed): error finding address for *dotasks.LoadBalancer {"Name":"api-e2e-test-do-k8s-local","ID":"d0fa1b62-29b3-4533-ad6b-5dd13d884d96","Lifecycle":"Sync","Region":"nyc1","DropletTag":"KubernetesCluster-Master:e2e-test-do-k8s-local","IPAddress":"","ForAPIServer":true}: IP Address is still empty.
W0718 19:42:40.834153   11766 executor.go:139] error running task "BootstrapScript/master-nyc1-2" (7m26s remaining to succeed): error finding address for *dotasks.LoadBalancer {"Name":"api-e2e-test-do-k8s-local","ID":"d0fa1b62-29b3-4533-ad6b-5dd13d884d96","Lifecycle":"Sync","Region":"nyc1","DropletTag":"KubernetesCluster-Master:e2e-test-do-k8s-local","IPAddress":"","ForAPIServer":true}: IP Address is still empty.
W0718 19:42:40.834175   11766 executor.go:139] error running task "BootstrapScript/nodes-nyc1" (7m36s remaining to succeed): error finding address for *dotasks.LoadBalancer {"Name":"api-e2e-test-do-k8s-local","ID":"d0fa1b62-29b3-4533-ad6b-5dd13d884d96","Lifecycle":"Sync","Region":"nyc1","DropletTag":"KubernetesCluster-Master:e2e-test-do-k8s-local","IPAddress":"","ForAPIServer":true}: IP Address is still empty.
I0718 19:42:40.834183   11766 executor.go:155] No progress made, sleeping before retrying 4 task(s)
I0718 19:42:50.834466   11766 executor.go:111] Tasks: 45 done / 57 total; 4 can run
I0718 19:42:51.213695   11766 executor.go:111] Tasks: 49 done / 57 total; 8 can run
I0718 19:42:53.121534   11766 executor.go:111] Tasks: 57 done / 57 total; 0 can run
I0718 19:42:53.243847   11766 update_cluster.go:326] Exporting kubeconfig for cluster
kOps has set your kubectl context to e2e-test-do.k8s.local
... skipping 10 lines ...
I0718 19:42:53.716211    4255 up.go:181] /home/prow/go/src/k8s.io/kops/bazel-bin/cmd/kops/linux-amd64/kops validate cluster --name e2e-test-do.k8s.local --count 10 --wait 20m0s
I0718 19:42:53.734059   11788 featureflag.go:167] FeatureFlag "SpecOverrideFlag"=true
I0718 19:42:53.734182   11788 featureflag.go:167] FeatureFlag "AlphaAllowGCE"=true
I0718 19:42:53.777342   11788 s3context.go:92] Found S3_ENDPOINT="sfo3.digitaloceanspaces.com", using as non-AWS S3 backend
Validating cluster e2e-test-do.k8s.local

W0718 19:43:04.820531   11788 validate_cluster.go:184] (will retry): unexpected error during validation: error listing nodes: an error on the server ("") has prevented the request from succeeding (get nodes)
W0718 19:43:25.432295   11788 validate_cluster.go:184] (will retry): unexpected error during validation: error listing nodes: an error on the server ("") has prevented the request from succeeding (get nodes)
W0718 19:43:46.042350   11788 validate_cluster.go:184] (will retry): unexpected error during validation: error listing nodes: an error on the server ("") has prevented the request from succeeding (get nodes)
W0718 19:44:09.410150   11788 validate_cluster.go:184] (will retry): unexpected error during validation: error listing nodes: an error on the server ("") has prevented the request from succeeding (get nodes)
W0718 19:44:30.023375   11788 validate_cluster.go:184] (will retry): unexpected error during validation: error listing nodes: an error on the server ("") has prevented the request from succeeding (get nodes)
W0718 19:44:50.635925   11788 validate_cluster.go:184] (will retry): unexpected error during validation: error listing nodes: an error on the server ("") has prevented the request from succeeding (get nodes)
W0718 19:45:11.245986   11788 validate_cluster.go:184] (will retry): unexpected error during validation: error listing nodes: an error on the server ("") has prevented the request from succeeding (get nodes)
W0718 19:45:31.862825   11788 validate_cluster.go:184] (will retry): unexpected error during validation: error listing nodes: an error on the server ("") has prevented the request from succeeding (get nodes)
W0718 19:45:52.482332   11788 validate_cluster.go:184] (will retry): unexpected error during validation: error listing nodes: an error on the server ("") has prevented the request from succeeding (get nodes)
W0718 19:46:13.097467   11788 validate_cluster.go:184] (will retry): unexpected error during validation: error listing nodes: an error on the server ("") has prevented the request from succeeding (get nodes)
W0718 19:46:33.711687   11788 validate_cluster.go:184] (will retry): unexpected error during validation: error listing nodes: an error on the server ("") has prevented the request from succeeding (get nodes)
W0718 19:46:54.327156   11788 validate_cluster.go:184] (will retry): unexpected error during validation: error listing nodes: an error on the server ("") has prevented the request from succeeding (get nodes)
W0718 19:47:14.941162   11788 validate_cluster.go:184] (will retry): unexpected error during validation: error listing nodes: an error on the server ("") has prevented the request from succeeding (get nodes)
W0718 19:47:35.557180   11788 validate_cluster.go:184] (will retry): unexpected error during validation: error listing nodes: an error on the server ("") has prevented the request from succeeding (get nodes)
W0718 19:47:56.176049   11788 validate_cluster.go:184] (will retry): unexpected error during validation: error listing nodes: an error on the server ("") has prevented the request from succeeding (get nodes)
W0718 19:48:16.788753   11788 validate_cluster.go:184] (will retry): unexpected error during validation: error listing nodes: an error on the server ("") has prevented the request from succeeding (get nodes)
INSTANCE GROUPS
NAME		ROLE	MACHINETYPE	MIN	MAX	SUBNETS
master-nyc1-1	Master	c2-16vcpu-32gb	1	1	nyc1
master-nyc1-2	Master	c2-16vcpu-32gb	1	1	nyc1
master-nyc1-3	Master	c2-16vcpu-32gb	1	1	nyc1
nodes-nyc1	Node	c2-16vcpu-32gb	2	2	nyc1
... skipping 545 lines ...
[sig-storage] In-tree Volumes
/workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/storage/utils/framework.go:23
  [Driver: gcepd]
  /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/storage/in_tree_volumes.go:58
    [Testpattern: Dynamic PV (immediate binding)] topology
    /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/storage/framework/testsuite.go:50
      should fail to schedule a pod which has topologies that conflict with AllowedTopologies [BeforeEach]
      /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/storage/testsuites/topology.go:192

      Only supported for providers [gce gke] (not skeleton)

      /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/storage/drivers/in_tree.go:1301
------------------------------
... skipping 468 lines ...
  /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:186
Jul 18 19:50:36.901: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready
STEP: Destroying namespace "request-timeout-9587" for this suite.

•
------------------------------
{"msg":"PASSED [sig-api-machinery] Server request timeout should return HTTP status code 400 if the user specifies an invalid timeout in the request URL","total":-1,"completed":1,"skipped":7,"failed":0}

SSS
------------------------------
[BeforeEach] [Testpattern: Pre-provisioned PV (default fs)] subPath
  /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/storage/framework/testsuite.go:51
Jul 18 19:50:37.013: INFO: Driver emptydir doesn't support PreprovisionedPV -- skipping
... skipping 28 lines ...
[sig-storage] In-tree Volumes
/workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/storage/utils/framework.go:23
  [Driver: hostPath]
  /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/storage/in_tree_volumes.go:58
    [Testpattern: Dynamic PV (immediate binding)] topology
    /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/storage/framework/testsuite.go:50
      should fail to schedule a pod which has topologies that conflict with AllowedTopologies [BeforeEach]
      /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/storage/testsuites/topology.go:192

      Driver hostPath doesn't support DynamicPV -- skipping

      /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/storage/framework/testsuite.go:116
------------------------------
... skipping 101 lines ...
STEP: Building a namespace api object, basename volume-provisioning
W0718 19:50:37.265231   12368 warnings.go:70] policy/v1beta1 PodSecurityPolicy is deprecated in v1.21+, unavailable in v1.25+
Jul 18 19:50:37.265: INFO: No PodSecurityPolicies found; assuming PodSecurityPolicy is disabled.
STEP: Waiting for a default service account to be provisioned in namespace
[BeforeEach] [sig-storage] Dynamic Provisioning
  /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/storage/volume_provisioning.go:146
[It] should report an error and create no PV
  /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/storage/volume_provisioning.go:825
Jul 18 19:50:37.326: INFO: Only supported for providers [aws] (not skeleton)
[AfterEach] [sig-storage] Dynamic Provisioning
  /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:186
Jul 18 19:50:37.327: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready
STEP: Destroying namespace "volume-provisioning-8478" for this suite.


S [SKIPPING] [2.021 seconds]
[sig-storage] Dynamic Provisioning
/workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/storage/utils/framework.go:23
  Invalid AWS KMS key
  /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/storage/volume_provisioning.go:824
    should report an error and create no PV [It]
    /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/storage/volume_provisioning.go:825

    Only supported for providers [aws] (not skeleton)

    /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/storage/volume_provisioning.go:826
------------------------------
... skipping 41 lines ...
  /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:186
Jul 18 19:50:37.339: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready
STEP: Destroying namespace "disruption-8908" for this suite.

•S
------------------------------
{"msg":"PASSED [sig-apps] DisruptionController should create a PodDisruptionBudget [Conformance]","total":-1,"completed":1,"skipped":11,"failed":0}

SSSSSSSSSSS
------------------------------
[BeforeEach] [sig-node] Security Context
  /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:185
STEP: Creating a kubernetes client
... skipping 10 lines ...
  /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:186
Jul 18 19:50:37.588: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready
STEP: Destroying namespace "security-context-test-13" for this suite.

•
------------------------------
{"msg":"PASSED [sig-node] Security Context When creating a container with runAsNonRoot should not run without a specified user ID","total":-1,"completed":1,"skipped":4,"failed":0}

S
------------------------------
[BeforeEach] [Testpattern: Dynamic PV (block volmode)] volumes
  /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/storage/framework/testsuite.go:51
Jul 18 19:50:37.734: INFO: Driver emptydir doesn't support DynamicPV -- skipping
... skipping 94 lines ...
  /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:186
Jul 18 19:50:38.555: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready
STEP: Destroying namespace "ingress-9159" for this suite.

•
------------------------------
{"msg":"PASSED [sig-network] Ingress API should support creating Ingress API operations [Conformance]","total":-1,"completed":1,"skipped":22,"failed":0}

S
------------------------------
[BeforeEach] [Testpattern: Dynamic PV (default fs)] fsgroupchangepolicy
  /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/storage/framework/testsuite.go:51
Jul 18 19:50:38.665: INFO: Driver local doesn't support DynamicPV -- skipping
... skipping 130 lines ...
  /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:186
Jul 18 19:50:39.653: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready
STEP: Destroying namespace "nettest-2145" for this suite.

•
------------------------------
{"msg":"PASSED [sig-network] Networking should provide unchanging, static URL paths for kubernetes api services","total":-1,"completed":2,"skipped":16,"failed":0}

SS
------------------------------
[BeforeEach] [sig-node] Kubelet
  /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:185
STEP: Creating a kubernetes client
... skipping 22 lines ...
/workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/common/node/framework.go:23
  when scheduling a read only busybox container
  /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/common/node/kubelet.go:188
    should not write to root filesystem [LinuxOnly] [NodeConformance] [Conformance]
    /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:630
------------------------------
{"msg":"PASSED [sig-node] Kubelet when scheduling a read only busybox container should not write to root filesystem [LinuxOnly] [NodeConformance] [Conformance]","total":-1,"completed":1,"skipped":1,"failed":0}

SSS
------------------------------
[BeforeEach] [sig-node] Security Context
  /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:185
STEP: Creating a kubernetes client
... skipping 2 lines ...
W0718 19:50:35.306865   12369 warnings.go:70] policy/v1beta1 PodSecurityPolicy is deprecated in v1.21+, unavailable in v1.25+
Jul 18 19:50:35.308: INFO: No PodSecurityPolicies found; assuming PodSecurityPolicy is disabled.
STEP: Waiting for a default service account to be provisioned in namespace
[It] should support container.SecurityContext.RunAsUser [LinuxOnly]
  /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/node/security_context.go:109
STEP: Creating a pod to test pod.Spec.SecurityContext.RunAsUser
Jul 18 19:50:35.422: INFO: Waiting up to 5m0s for pod "security-context-c0737702-04d5-4025-a63a-78e74f0284bb" in namespace "security-context-3444" to be "Succeeded or Failed"
Jul 18 19:50:35.452: INFO: Pod "security-context-c0737702-04d5-4025-a63a-78e74f0284bb": Phase="Pending", Reason="", readiness=false. Elapsed: 30.456166ms
Jul 18 19:50:37.483: INFO: Pod "security-context-c0737702-04d5-4025-a63a-78e74f0284bb": Phase="Pending", Reason="", readiness=false. Elapsed: 2.061114789s
Jul 18 19:50:39.746: INFO: Pod "security-context-c0737702-04d5-4025-a63a-78e74f0284bb": Phase="Pending", Reason="", readiness=false. Elapsed: 4.323858129s
Jul 18 19:50:41.782: INFO: Pod "security-context-c0737702-04d5-4025-a63a-78e74f0284bb": Phase="Pending", Reason="", readiness=false. Elapsed: 6.359694378s
Jul 18 19:50:43.814: INFO: Pod "security-context-c0737702-04d5-4025-a63a-78e74f0284bb": Phase="Succeeded", Reason="", readiness=false. Elapsed: 8.391840589s
STEP: Saw pod success
Jul 18 19:50:43.814: INFO: Pod "security-context-c0737702-04d5-4025-a63a-78e74f0284bb" satisfied condition "Succeeded or Failed"
Jul 18 19:50:43.844: INFO: Trying to get logs from node 10.116.0.5 pod security-context-c0737702-04d5-4025-a63a-78e74f0284bb container test-container: <nil>
STEP: delete the pod
Jul 18 19:50:44.294: INFO: Waiting for pod security-context-c0737702-04d5-4025-a63a-78e74f0284bb to disappear
Jul 18 19:50:44.324: INFO: Pod security-context-c0737702-04d5-4025-a63a-78e74f0284bb no longer exists
[AfterEach] [sig-node] Security Context
  /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:186
... skipping 4 lines ...
• [SLOW TEST:9.250 seconds]
[sig-node] Security Context
/workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/node/framework.go:23
  should support container.SecurityContext.RunAsUser [LinuxOnly]
  /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/node/security_context.go:109
------------------------------
{"msg":"PASSED [sig-node] Security Context should support container.SecurityContext.RunAsUser [LinuxOnly]","total":-1,"completed":1,"skipped":1,"failed":0}

SSS
------------------------------
[BeforeEach] [Testpattern: Dynamic PV (default fs)] subPath
  /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/storage/framework/testsuite.go:51
Jul 18 19:50:44.438: INFO: Driver local doesn't support DynamicPV -- skipping
... skipping 48 lines ...
STEP: Waiting for a default service account to be provisioned in namespace
[BeforeEach] [sig-storage] Projected downwardAPI
  /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/common/storage/projected_downwardapi.go:41
[It] should set mode on item file [LinuxOnly] [NodeConformance] [Conformance]
  /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:630
STEP: Creating a pod to test downward API volume plugin
Jul 18 19:50:35.643: INFO: Waiting up to 5m0s for pod "downwardapi-volume-cac78fa6-e75d-4f0c-a173-fbaa90de750c" in namespace "projected-8641" to be "Succeeded or Failed"
Jul 18 19:50:35.674: INFO: Pod "downwardapi-volume-cac78fa6-e75d-4f0c-a173-fbaa90de750c": Phase="Pending", Reason="", readiness=false. Elapsed: 30.994364ms
Jul 18 19:50:37.712: INFO: Pod "downwardapi-volume-cac78fa6-e75d-4f0c-a173-fbaa90de750c": Phase="Pending", Reason="", readiness=false. Elapsed: 2.06846213s
Jul 18 19:50:39.746: INFO: Pod "downwardapi-volume-cac78fa6-e75d-4f0c-a173-fbaa90de750c": Phase="Pending", Reason="", readiness=false. Elapsed: 4.103138489s
Jul 18 19:50:41.782: INFO: Pod "downwardapi-volume-cac78fa6-e75d-4f0c-a173-fbaa90de750c": Phase="Pending", Reason="", readiness=false. Elapsed: 6.138708923s
Jul 18 19:50:43.814: INFO: Pod "downwardapi-volume-cac78fa6-e75d-4f0c-a173-fbaa90de750c": Phase="Pending", Reason="", readiness=false. Elapsed: 8.170573998s
Jul 18 19:50:45.845: INFO: Pod "downwardapi-volume-cac78fa6-e75d-4f0c-a173-fbaa90de750c": Phase="Succeeded", Reason="", readiness=false. Elapsed: 10.201577597s
STEP: Saw pod success
Jul 18 19:50:45.845: INFO: Pod "downwardapi-volume-cac78fa6-e75d-4f0c-a173-fbaa90de750c" satisfied condition "Succeeded or Failed"
Jul 18 19:50:45.875: INFO: Trying to get logs from node 10.116.0.5 pod downwardapi-volume-cac78fa6-e75d-4f0c-a173-fbaa90de750c container client-container: <nil>
STEP: delete the pod
Jul 18 19:50:45.953: INFO: Waiting for pod downwardapi-volume-cac78fa6-e75d-4f0c-a173-fbaa90de750c to disappear
Jul 18 19:50:45.983: INFO: Pod downwardapi-volume-cac78fa6-e75d-4f0c-a173-fbaa90de750c no longer exists
[AfterEach] [sig-storage] Projected downwardAPI
  /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:186
... skipping 15 lines ...
W0718 19:50:35.432134   12499 warnings.go:70] policy/v1beta1 PodSecurityPolicy is deprecated in v1.21+, unavailable in v1.25+
Jul 18 19:50:35.432: INFO: No PodSecurityPolicies found; assuming PodSecurityPolicy is disabled.
STEP: Waiting for a default service account to be provisioned in namespace
[It] should support (non-root,0666,tmpfs) [LinuxOnly] [NodeConformance] [Conformance]
  /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:630
STEP: Creating a pod to test emptydir 0666 on tmpfs
Jul 18 19:50:35.556: INFO: Waiting up to 5m0s for pod "pod-cf28e437-7627-4e93-8264-21cdd359e9ad" in namespace "emptydir-4973" to be "Succeeded or Failed"
Jul 18 19:50:35.598: INFO: Pod "pod-cf28e437-7627-4e93-8264-21cdd359e9ad": Phase="Pending", Reason="", readiness=false. Elapsed: 41.939399ms
Jul 18 19:50:37.630: INFO: Pod "pod-cf28e437-7627-4e93-8264-21cdd359e9ad": Phase="Pending", Reason="", readiness=false. Elapsed: 2.074566361s
Jul 18 19:50:39.746: INFO: Pod "pod-cf28e437-7627-4e93-8264-21cdd359e9ad": Phase="Pending", Reason="", readiness=false. Elapsed: 4.190077692s
Jul 18 19:50:41.777: INFO: Pod "pod-cf28e437-7627-4e93-8264-21cdd359e9ad": Phase="Pending", Reason="", readiness=false. Elapsed: 6.221792195s
Jul 18 19:50:43.810: INFO: Pod "pod-cf28e437-7627-4e93-8264-21cdd359e9ad": Phase="Pending", Reason="", readiness=false. Elapsed: 8.254264157s
Jul 18 19:50:45.843: INFO: Pod "pod-cf28e437-7627-4e93-8264-21cdd359e9ad": Phase="Succeeded", Reason="", readiness=false. Elapsed: 10.287689535s
STEP: Saw pod success
Jul 18 19:50:45.843: INFO: Pod "pod-cf28e437-7627-4e93-8264-21cdd359e9ad" satisfied condition "Succeeded or Failed"
Jul 18 19:50:45.875: INFO: Trying to get logs from node 10.116.0.7 pod pod-cf28e437-7627-4e93-8264-21cdd359e9ad container test-container: <nil>
STEP: delete the pod
Jul 18 19:50:45.957: INFO: Waiting for pod pod-cf28e437-7627-4e93-8264-21cdd359e9ad to disappear
Jul 18 19:50:45.987: INFO: Pod pod-cf28e437-7627-4e93-8264-21cdd359e9ad no longer exists
[AfterEach] [sig-storage] EmptyDir volumes
  /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:186
... skipping 4 lines ...
• [SLOW TEST:10.793 seconds]
[sig-storage] EmptyDir volumes
/workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/common/storage/framework.go:23
  should support (non-root,0666,tmpfs) [LinuxOnly] [NodeConformance] [Conformance]
  /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:630
------------------------------
{"msg":"PASSED [sig-storage] EmptyDir volumes should support (non-root,0666,tmpfs) [LinuxOnly] [NodeConformance] [Conformance]","total":-1,"completed":1,"skipped":2,"failed":0}

S
------------------------------
{"msg":"PASSED [sig-storage] Projected downwardAPI should set mode on item file [LinuxOnly] [NodeConformance] [Conformance]","total":-1,"completed":1,"skipped":2,"failed":0}
[BeforeEach] [Testpattern: Dynamic PV (default fs)] subPath
  /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/storage/framework/testsuite.go:51
Jul 18 19:50:46.089: INFO: Only supported for providers [vsphere] (not skeleton)
[AfterEach] [Testpattern: Dynamic PV (default fs)] subPath
  /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:186

... skipping 64 lines ...
W0718 19:50:36.060700   12449 warnings.go:70] policy/v1beta1 PodSecurityPolicy is deprecated in v1.21+, unavailable in v1.25+
Jul 18 19:50:36.060: INFO: No PodSecurityPolicies found; assuming PodSecurityPolicy is disabled.
STEP: Waiting for a default service account to be provisioned in namespace
[It] should support (root,0777,tmpfs) [LinuxOnly] [NodeConformance] [Conformance]
  /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:630
STEP: Creating a pod to test emptydir 0777 on tmpfs
Jul 18 19:50:36.155: INFO: Waiting up to 5m0s for pod "pod-a0345526-2357-4290-ad10-6cf740aa2852" in namespace "emptydir-4774" to be "Succeeded or Failed"
Jul 18 19:50:36.185: INFO: Pod "pod-a0345526-2357-4290-ad10-6cf740aa2852": Phase="Pending", Reason="", readiness=false. Elapsed: 30.258028ms
Jul 18 19:50:38.217: INFO: Pod "pod-a0345526-2357-4290-ad10-6cf740aa2852": Phase="Pending", Reason="", readiness=false. Elapsed: 2.061547192s
Jul 18 19:50:40.249: INFO: Pod "pod-a0345526-2357-4290-ad10-6cf740aa2852": Phase="Pending", Reason="", readiness=false. Elapsed: 4.093854046s
Jul 18 19:50:42.280: INFO: Pod "pod-a0345526-2357-4290-ad10-6cf740aa2852": Phase="Pending", Reason="", readiness=false. Elapsed: 6.125175016s
Jul 18 19:50:44.313: INFO: Pod "pod-a0345526-2357-4290-ad10-6cf740aa2852": Phase="Pending", Reason="", readiness=false. Elapsed: 8.157606664s
Jul 18 19:50:46.345: INFO: Pod "pod-a0345526-2357-4290-ad10-6cf740aa2852": Phase="Succeeded", Reason="", readiness=false. Elapsed: 10.190080575s
STEP: Saw pod success
Jul 18 19:50:46.345: INFO: Pod "pod-a0345526-2357-4290-ad10-6cf740aa2852" satisfied condition "Succeeded or Failed"
Jul 18 19:50:46.376: INFO: Trying to get logs from node 10.116.0.5 pod pod-a0345526-2357-4290-ad10-6cf740aa2852 container test-container: <nil>
STEP: delete the pod
Jul 18 19:50:46.453: INFO: Waiting for pod pod-a0345526-2357-4290-ad10-6cf740aa2852 to disappear
Jul 18 19:50:46.484: INFO: Pod pod-a0345526-2357-4290-ad10-6cf740aa2852 no longer exists
[AfterEach] [sig-storage] EmptyDir volumes
  /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:186
... skipping 4 lines ...
• [SLOW TEST:11.252 seconds]
[sig-storage] EmptyDir volumes
/workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/common/storage/framework.go:23
  should support (root,0777,tmpfs) [LinuxOnly] [NodeConformance] [Conformance]
  /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:630
------------------------------
{"msg":"PASSED [sig-storage] EmptyDir volumes should support (root,0777,tmpfs) [LinuxOnly] [NodeConformance] [Conformance]","total":-1,"completed":1,"skipped":2,"failed":0}

SSSSSS
------------------------------
[BeforeEach] [Testpattern: Pre-provisioned PV (ext3)] volumes
  /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/storage/framework/testsuite.go:51
Jul 18 19:50:46.606: INFO: Driver local doesn't support ext3 -- skipping
... skipping 74 lines ...
[It] should support existing single file [LinuxOnly]
  /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/storage/testsuites/subpath.go:219
Jul 18 19:50:39.808: INFO: In-tree plugin kubernetes.io/empty-dir is not migrated, not validating any metrics
Jul 18 19:50:39.808: INFO: Creating resource for inline volume
STEP: Creating pod pod-subpath-test-inlinevolume-gl6n
STEP: Creating a pod to test subpath
Jul 18 19:50:39.851: INFO: Waiting up to 5m0s for pod "pod-subpath-test-inlinevolume-gl6n" in namespace "provisioning-3025" to be "Succeeded or Failed"
Jul 18 19:50:39.888: INFO: Pod "pod-subpath-test-inlinevolume-gl6n": Phase="Pending", Reason="", readiness=false. Elapsed: 36.955413ms
Jul 18 19:50:41.919: INFO: Pod "pod-subpath-test-inlinevolume-gl6n": Phase="Pending", Reason="", readiness=false. Elapsed: 2.06859318s
Jul 18 19:50:43.951: INFO: Pod "pod-subpath-test-inlinevolume-gl6n": Phase="Pending", Reason="", readiness=false. Elapsed: 4.100344748s
Jul 18 19:50:45.982: INFO: Pod "pod-subpath-test-inlinevolume-gl6n": Phase="Pending", Reason="", readiness=false. Elapsed: 6.131704604s
Jul 18 19:50:48.015: INFO: Pod "pod-subpath-test-inlinevolume-gl6n": Phase="Succeeded", Reason="", readiness=false. Elapsed: 8.164571009s
STEP: Saw pod success
Jul 18 19:50:48.015: INFO: Pod "pod-subpath-test-inlinevolume-gl6n" satisfied condition "Succeeded or Failed"
Jul 18 19:50:48.047: INFO: Trying to get logs from node 10.116.0.5 pod pod-subpath-test-inlinevolume-gl6n container test-container-subpath-inlinevolume-gl6n: <nil>
STEP: delete the pod
Jul 18 19:50:48.126: INFO: Waiting for pod pod-subpath-test-inlinevolume-gl6n to disappear
Jul 18 19:50:48.156: INFO: Pod pod-subpath-test-inlinevolume-gl6n no longer exists
STEP: Deleting pod pod-subpath-test-inlinevolume-gl6n
Jul 18 19:50:48.156: INFO: Deleting pod "pod-subpath-test-inlinevolume-gl6n" in namespace "provisioning-3025"
... skipping 24 lines ...
Jul 18 19:50:36.111: INFO: No PodSecurityPolicies found; assuming PodSecurityPolicy is disabled.
STEP: Waiting for a default service account to be provisioned in namespace
[It] should be able to mount in a volume regardless of a different secret existing with same name in different namespace [NodeConformance] [Conformance]
  /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:630
STEP: Creating secret with name secret-test-eb9f3735-80ae-4e18-b8a6-1c0c25a919b8
STEP: Creating a pod to test consume secrets
Jul 18 19:50:38.087: INFO: Waiting up to 5m0s for pod "pod-secrets-3da71931-c67e-4d0c-b930-596578b9bad0" in namespace "secrets-872" to be "Succeeded or Failed"
Jul 18 19:50:38.119: INFO: Pod "pod-secrets-3da71931-c67e-4d0c-b930-596578b9bad0": Phase="Pending", Reason="", readiness=false. Elapsed: 31.250488ms
Jul 18 19:50:40.150: INFO: Pod "pod-secrets-3da71931-c67e-4d0c-b930-596578b9bad0": Phase="Pending", Reason="", readiness=false. Elapsed: 2.063030819s
Jul 18 19:50:42.183: INFO: Pod "pod-secrets-3da71931-c67e-4d0c-b930-596578b9bad0": Phase="Pending", Reason="", readiness=false. Elapsed: 4.095455412s
Jul 18 19:50:44.215: INFO: Pod "pod-secrets-3da71931-c67e-4d0c-b930-596578b9bad0": Phase="Pending", Reason="", readiness=false. Elapsed: 6.127370258s
Jul 18 19:50:46.249: INFO: Pod "pod-secrets-3da71931-c67e-4d0c-b930-596578b9bad0": Phase="Pending", Reason="", readiness=false. Elapsed: 8.16143771s
Jul 18 19:50:48.281: INFO: Pod "pod-secrets-3da71931-c67e-4d0c-b930-596578b9bad0": Phase="Succeeded", Reason="", readiness=false. Elapsed: 10.19344432s
STEP: Saw pod success
Jul 18 19:50:48.281: INFO: Pod "pod-secrets-3da71931-c67e-4d0c-b930-596578b9bad0" satisfied condition "Succeeded or Failed"
Jul 18 19:50:48.311: INFO: Trying to get logs from node 10.116.0.7 pod pod-secrets-3da71931-c67e-4d0c-b930-596578b9bad0 container secret-volume-test: <nil>
STEP: delete the pod
Jul 18 19:50:48.390: INFO: Waiting for pod pod-secrets-3da71931-c67e-4d0c-b930-596578b9bad0 to disappear
Jul 18 19:50:48.421: INFO: Pod pod-secrets-3da71931-c67e-4d0c-b930-596578b9bad0 no longer exists
[AfterEach] [sig-storage] Secrets
  /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:186
... skipping 5 lines ...
• [SLOW TEST:13.226 seconds]
[sig-storage] Secrets
/workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/common/storage/framework.go:23
  should be able to mount in a volume regardless of a different secret existing with same name in different namespace [NodeConformance] [Conformance]
  /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:630
------------------------------
{"msg":"PASSED [sig-storage] Secrets should be able to mount in a volume regardless of a different secret existing with same name in different namespace [NodeConformance] [Conformance]","total":-1,"completed":1,"skipped":2,"failed":0}

SSS
------------------------------
[BeforeEach] [sig-node] Secrets
  /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:185
STEP: Creating a kubernetes client
Jul 18 19:50:39.816: INFO: >>> kubeConfig: /root/.kube/config
STEP: Building a namespace api object, basename secrets
STEP: Waiting for a default service account to be provisioned in namespace
[It] should be consumable via the environment [NodeConformance] [Conformance]
  /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:630
STEP: creating secret secrets-1449/secret-test-b06d23ec-dca3-4565-b302-1663e3c79f60
STEP: Creating a pod to test consume secrets
Jul 18 19:50:40.167: INFO: Waiting up to 5m0s for pod "pod-configmaps-04932edc-b69c-42ca-a082-2d9417c847cf" in namespace "secrets-1449" to be "Succeeded or Failed"
Jul 18 19:50:40.203: INFO: Pod "pod-configmaps-04932edc-b69c-42ca-a082-2d9417c847cf": Phase="Pending", Reason="", readiness=false. Elapsed: 35.437605ms
Jul 18 19:50:42.235: INFO: Pod "pod-configmaps-04932edc-b69c-42ca-a082-2d9417c847cf": Phase="Pending", Reason="", readiness=false. Elapsed: 2.067193627s
Jul 18 19:50:44.267: INFO: Pod "pod-configmaps-04932edc-b69c-42ca-a082-2d9417c847cf": Phase="Pending", Reason="", readiness=false. Elapsed: 4.099535066s
Jul 18 19:50:46.301: INFO: Pod "pod-configmaps-04932edc-b69c-42ca-a082-2d9417c847cf": Phase="Pending", Reason="", readiness=false. Elapsed: 6.1332215s
Jul 18 19:50:48.332: INFO: Pod "pod-configmaps-04932edc-b69c-42ca-a082-2d9417c847cf": Phase="Succeeded", Reason="", readiness=false. Elapsed: 8.16445263s
STEP: Saw pod success
Jul 18 19:50:48.332: INFO: Pod "pod-configmaps-04932edc-b69c-42ca-a082-2d9417c847cf" satisfied condition "Succeeded or Failed"
Jul 18 19:50:48.363: INFO: Trying to get logs from node 10.116.0.5 pod pod-configmaps-04932edc-b69c-42ca-a082-2d9417c847cf container env-test: <nil>
STEP: delete the pod
Jul 18 19:50:48.449: INFO: Waiting for pod pod-configmaps-04932edc-b69c-42ca-a082-2d9417c847cf to disappear
Jul 18 19:50:48.481: INFO: Pod pod-configmaps-04932edc-b69c-42ca-a082-2d9417c847cf no longer exists
[AfterEach] [sig-node] Secrets
  /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:186
... skipping 4 lines ...
• [SLOW TEST:8.762 seconds]
[sig-node] Secrets
/workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/common/node/framework.go:23
  should be consumable via the environment [NodeConformance] [Conformance]
  /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:630
------------------------------
{"msg":"PASSED [sig-node] Secrets should be consumable via the environment [NodeConformance] [Conformance]","total":-1,"completed":3,"skipped":18,"failed":0}

SSSS
------------------------------
[BeforeEach] [Testpattern: Dynamic PV (ext4)] volumes
  /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/storage/framework/testsuite.go:51
Jul 18 19:50:48.601: INFO: Driver csi-hostpath doesn't support ext4 -- skipping
... skipping 58 lines ...
      /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/storage/testsuites/volumes.go:196

      Driver emptydir doesn't support DynamicPV -- skipping

      /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/storage/framework/testsuite.go:116
------------------------------
{"msg":"PASSED [sig-instrumentation] MetricsGrabber should grab all metrics from a Scheduler.","total":-1,"completed":2,"skipped":15,"failed":0}
[BeforeEach] [sig-node] Pods Extended
  /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:185
STEP: Creating a kubernetes client
Jul 18 19:50:38.568: INFO: >>> kubeConfig: /root/.kube/config
STEP: Building a namespace api object, basename pods
STEP: Waiting for a default service account to be provisioned in namespace
[BeforeEach] Pod Container lifecycle
  /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/node/pods.go:446
[It] should not create extra sandbox if all containers are done
  /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/node/pods.go:450
STEP: creating the pod that should always exit 0
STEP: submitting the pod to kubernetes
Jul 18 19:50:39.470: INFO: Waiting up to 5m0s for pod "pod-always-succeed2731a486-04d5-4d60-8c32-e9a2250a441b" in namespace "pods-6322" to be "Succeeded or Failed"
Jul 18 19:50:39.501: INFO: Pod "pod-always-succeed2731a486-04d5-4d60-8c32-e9a2250a441b": Phase="Pending", Reason="", readiness=false. Elapsed: 30.802412ms
Jul 18 19:50:41.533: INFO: Pod "pod-always-succeed2731a486-04d5-4d60-8c32-e9a2250a441b": Phase="Pending", Reason="", readiness=false. Elapsed: 2.062954631s
Jul 18 19:50:43.564: INFO: Pod "pod-always-succeed2731a486-04d5-4d60-8c32-e9a2250a441b": Phase="Pending", Reason="", readiness=false. Elapsed: 4.093901782s
Jul 18 19:50:45.596: INFO: Pod "pod-always-succeed2731a486-04d5-4d60-8c32-e9a2250a441b": Phase="Pending", Reason="", readiness=false. Elapsed: 6.125823674s
Jul 18 19:50:47.629: INFO: Pod "pod-always-succeed2731a486-04d5-4d60-8c32-e9a2250a441b": Phase="Succeeded", Reason="", readiness=false. Elapsed: 8.158827688s
STEP: Saw pod success
Jul 18 19:50:47.629: INFO: Pod "pod-always-succeed2731a486-04d5-4d60-8c32-e9a2250a441b" satisfied condition "Succeeded or Failed"
STEP: Getting events about the pod
STEP: Checking events about the pod
STEP: deleting the pod
[AfterEach] [sig-node] Pods Extended
  /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:186
Jul 18 19:50:49.700: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready
... skipping 5 lines ...
/workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/node/framework.go:23
  Pod Container lifecycle
  /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/node/pods.go:444
    should not create extra sandbox if all containers are done
    /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/node/pods.go:450
------------------------------
{"msg":"PASSED [sig-node] Pods Extended Pod Container lifecycle should not create extra sandbox if all containers are done","total":-1,"completed":3,"skipped":15,"failed":0}
[BeforeEach] [Testpattern: Inline-volume (default fs)] subPath
  /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/storage/framework/testsuite.go:51
[BeforeEach] [Testpattern: Inline-volume (default fs)] subPath
  /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:185
STEP: Creating a kubernetes client
Jul 18 19:50:49.804: INFO: >>> kubeConfig: /root/.kube/config
... skipping 133 lines ...
[It] should support existing single file [LinuxOnly]
  /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/storage/testsuites/subpath.go:219
Jul 18 19:50:37.668: INFO: In-tree plugin kubernetes.io/host-path is not migrated, not validating any metrics
Jul 18 19:50:37.727: INFO: Creating resource for inline volume
STEP: Creating pod pod-subpath-test-inlinevolume-7nhb
STEP: Creating a pod to test subpath
Jul 18 19:50:37.763: INFO: Waiting up to 5m0s for pod "pod-subpath-test-inlinevolume-7nhb" in namespace "provisioning-3263" to be "Succeeded or Failed"
Jul 18 19:50:37.794: INFO: Pod "pod-subpath-test-inlinevolume-7nhb": Phase="Pending", Reason="", readiness=false. Elapsed: 31.038831ms
Jul 18 19:50:39.826: INFO: Pod "pod-subpath-test-inlinevolume-7nhb": Phase="Pending", Reason="", readiness=false. Elapsed: 2.062853961s
Jul 18 19:50:41.859: INFO: Pod "pod-subpath-test-inlinevolume-7nhb": Phase="Pending", Reason="", readiness=false. Elapsed: 4.095641822s
Jul 18 19:50:43.890: INFO: Pod "pod-subpath-test-inlinevolume-7nhb": Phase="Pending", Reason="", readiness=false. Elapsed: 6.127105132s
Jul 18 19:50:45.923: INFO: Pod "pod-subpath-test-inlinevolume-7nhb": Phase="Pending", Reason="", readiness=false. Elapsed: 8.159766561s
Jul 18 19:50:47.955: INFO: Pod "pod-subpath-test-inlinevolume-7nhb": Phase="Pending", Reason="", readiness=false. Elapsed: 10.192040509s
Jul 18 19:50:49.986: INFO: Pod "pod-subpath-test-inlinevolume-7nhb": Phase="Pending", Reason="", readiness=false. Elapsed: 12.222969856s
Jul 18 19:50:52.018: INFO: Pod "pod-subpath-test-inlinevolume-7nhb": Phase="Succeeded", Reason="", readiness=false. Elapsed: 14.255249519s
STEP: Saw pod success
Jul 18 19:50:52.019: INFO: Pod "pod-subpath-test-inlinevolume-7nhb" satisfied condition "Succeeded or Failed"
Jul 18 19:50:52.049: INFO: Trying to get logs from node 10.116.0.5 pod pod-subpath-test-inlinevolume-7nhb container test-container-subpath-inlinevolume-7nhb: <nil>
STEP: delete the pod
Jul 18 19:50:52.124: INFO: Waiting for pod pod-subpath-test-inlinevolume-7nhb to disappear
Jul 18 19:50:52.156: INFO: Pod pod-subpath-test-inlinevolume-7nhb no longer exists
STEP: Deleting pod pod-subpath-test-inlinevolume-7nhb
Jul 18 19:50:52.156: INFO: Deleting pod "pod-subpath-test-inlinevolume-7nhb" in namespace "provisioning-3263"
... skipping 12 lines ...
  /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/storage/in_tree_volumes.go:58
    [Testpattern: Inline-volume (default fs)] subPath
    /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/storage/framework/testsuite.go:50
      should support existing single file [LinuxOnly]
      /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/storage/testsuites/subpath.go:219
------------------------------
{"msg":"PASSED [sig-storage] In-tree Volumes [Driver: hostPath] [Testpattern: Inline-volume (default fs)] subPath should support existing single file [LinuxOnly]","total":-1,"completed":1,"skipped":26,"failed":0}
[BeforeEach] [Testpattern: Pre-provisioned PV (default fs)] subPath
  /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/storage/framework/testsuite.go:51
Jul 18 19:50:52.321: INFO: Only supported for providers [gce gke] (not skeleton)
[AfterEach] [Testpattern: Pre-provisioned PV (default fs)] subPath
  /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:186

... skipping 192 lines ...
• [SLOW TEST:7.008 seconds]
[sig-api-machinery] Discovery
/workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/apimachinery/framework.go:23
  Custom resource should have storage version hash
  /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/apimachinery/discovery.go:48
------------------------------
{"msg":"PASSED [sig-api-machinery] Discovery Custom resource should have storage version hash","total":-1,"completed":2,"skipped":9,"failed":0}
[BeforeEach] [Testpattern: Pre-provisioned PV (default fs)] subPath
  /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/storage/framework/testsuite.go:51
Jul 18 19:50:53.140: INFO: Only supported for providers [gce gke] (not skeleton)
[AfterEach] [Testpattern: Pre-provisioned PV (default fs)] subPath
  /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:186

... skipping 22 lines ...
STEP: Waiting for a default service account to be provisioned in namespace
[It] should project all components that make up the projection API [Projection][NodeConformance] [Conformance]
  /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:630
STEP: Creating configMap with name configmap-projected-all-test-volume-ce441658-ba74-420e-8b3e-149872133332
STEP: Creating secret with name secret-projected-all-test-volume-8170bb5d-bc79-4dc3-a297-1e8bdc9b2a4b
STEP: Creating a pod to test Check all projections for projected volume plugin
Jul 18 19:50:35.642: INFO: Waiting up to 5m0s for pod "projected-volume-3df6bc1c-0747-4e8a-8040-19b4cea6e486" in namespace "projected-954" to be "Succeeded or Failed"
Jul 18 19:50:35.673: INFO: Pod "projected-volume-3df6bc1c-0747-4e8a-8040-19b4cea6e486": Phase="Pending", Reason="", readiness=false. Elapsed: 30.926838ms
Jul 18 19:50:37.711: INFO: Pod "projected-volume-3df6bc1c-0747-4e8a-8040-19b4cea6e486": Phase="Pending", Reason="", readiness=false. Elapsed: 2.068968305s
Jul 18 19:50:39.746: INFO: Pod "projected-volume-3df6bc1c-0747-4e8a-8040-19b4cea6e486": Phase="Pending", Reason="", readiness=false. Elapsed: 4.103901781s
Jul 18 19:50:41.782: INFO: Pod "projected-volume-3df6bc1c-0747-4e8a-8040-19b4cea6e486": Phase="Pending", Reason="", readiness=false. Elapsed: 6.139229359s
Jul 18 19:50:43.814: INFO: Pod "projected-volume-3df6bc1c-0747-4e8a-8040-19b4cea6e486": Phase="Pending", Reason="", readiness=false. Elapsed: 8.171363934s
Jul 18 19:50:45.845: INFO: Pod "projected-volume-3df6bc1c-0747-4e8a-8040-19b4cea6e486": Phase="Pending", Reason="", readiness=false. Elapsed: 10.202630222s
Jul 18 19:50:47.881: INFO: Pod "projected-volume-3df6bc1c-0747-4e8a-8040-19b4cea6e486": Phase="Pending", Reason="", readiness=false. Elapsed: 12.238315454s
Jul 18 19:50:49.937: INFO: Pod "projected-volume-3df6bc1c-0747-4e8a-8040-19b4cea6e486": Phase="Pending", Reason="", readiness=false. Elapsed: 14.294912947s
Jul 18 19:50:51.970: INFO: Pod "projected-volume-3df6bc1c-0747-4e8a-8040-19b4cea6e486": Phase="Pending", Reason="", readiness=false. Elapsed: 16.327215324s
Jul 18 19:50:54.002: INFO: Pod "projected-volume-3df6bc1c-0747-4e8a-8040-19b4cea6e486": Phase="Succeeded", Reason="", readiness=false. Elapsed: 18.359238091s
STEP: Saw pod success
Jul 18 19:50:54.002: INFO: Pod "projected-volume-3df6bc1c-0747-4e8a-8040-19b4cea6e486" satisfied condition "Succeeded or Failed"
Jul 18 19:50:54.035: INFO: Trying to get logs from node 10.116.0.5 pod projected-volume-3df6bc1c-0747-4e8a-8040-19b4cea6e486 container projected-all-volume-test: <nil>
STEP: delete the pod
Jul 18 19:50:54.179: INFO: Waiting for pod projected-volume-3df6bc1c-0747-4e8a-8040-19b4cea6e486 to disappear
Jul 18 19:50:54.210: INFO: Pod projected-volume-3df6bc1c-0747-4e8a-8040-19b4cea6e486 no longer exists
[AfterEach] [sig-storage] Projected combined
  /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:186
... skipping 4 lines ...
• [SLOW TEST:19.017 seconds]
[sig-storage] Projected combined
/workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/common/storage/framework.go:23
  should project all components that make up the projection API [Projection][NodeConformance] [Conformance]
  /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:630
------------------------------
{"msg":"PASSED [sig-storage] Projected combined should project all components that make up the projection API [Projection][NodeConformance] [Conformance]","total":-1,"completed":1,"skipped":1,"failed":0}

S
------------------------------
[BeforeEach] [sig-storage] Volumes
  /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:185
STEP: Creating a kubernetes client
... skipping 29 lines ...
Jul 18 19:50:37.465: INFO: >>> kubeConfig: /root/.kube/config
STEP: Building a namespace api object, basename downward-api
STEP: Waiting for a default service account to be provisioned in namespace
[It] should provide pod UID as env vars [NodeConformance] [Conformance]
  /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:630
STEP: Creating a pod to test downward api env vars
Jul 18 19:50:39.116: INFO: Waiting up to 5m0s for pod "downward-api-e8b45931-a5d6-4ec8-9f2b-47be44a5d33a" in namespace "downward-api-7634" to be "Succeeded or Failed"
Jul 18 19:50:39.145: INFO: Pod "downward-api-e8b45931-a5d6-4ec8-9f2b-47be44a5d33a": Phase="Pending", Reason="", readiness=false. Elapsed: 29.72331ms
Jul 18 19:50:41.177: INFO: Pod "downward-api-e8b45931-a5d6-4ec8-9f2b-47be44a5d33a": Phase="Pending", Reason="", readiness=false. Elapsed: 2.061000233s
Jul 18 19:50:43.209: INFO: Pod "downward-api-e8b45931-a5d6-4ec8-9f2b-47be44a5d33a": Phase="Pending", Reason="", readiness=false. Elapsed: 4.093728389s
Jul 18 19:50:45.242: INFO: Pod "downward-api-e8b45931-a5d6-4ec8-9f2b-47be44a5d33a": Phase="Pending", Reason="", readiness=false. Elapsed: 6.126223876s
Jul 18 19:50:47.274: INFO: Pod "downward-api-e8b45931-a5d6-4ec8-9f2b-47be44a5d33a": Phase="Pending", Reason="", readiness=false. Elapsed: 8.158642051s
Jul 18 19:50:49.306: INFO: Pod "downward-api-e8b45931-a5d6-4ec8-9f2b-47be44a5d33a": Phase="Pending", Reason="", readiness=false. Elapsed: 10.189865114s
Jul 18 19:50:51.337: INFO: Pod "downward-api-e8b45931-a5d6-4ec8-9f2b-47be44a5d33a": Phase="Pending", Reason="", readiness=false. Elapsed: 12.221494338s
Jul 18 19:50:53.370: INFO: Pod "downward-api-e8b45931-a5d6-4ec8-9f2b-47be44a5d33a": Phase="Pending", Reason="", readiness=false. Elapsed: 14.25391867s
Jul 18 19:50:55.401: INFO: Pod "downward-api-e8b45931-a5d6-4ec8-9f2b-47be44a5d33a": Phase="Succeeded", Reason="", readiness=false. Elapsed: 16.285251985s
STEP: Saw pod success
Jul 18 19:50:55.401: INFO: Pod "downward-api-e8b45931-a5d6-4ec8-9f2b-47be44a5d33a" satisfied condition "Succeeded or Failed"
Jul 18 19:50:55.433: INFO: Trying to get logs from node 10.116.0.5 pod downward-api-e8b45931-a5d6-4ec8-9f2b-47be44a5d33a container dapi-container: <nil>
STEP: delete the pod
Jul 18 19:50:55.515: INFO: Waiting for pod downward-api-e8b45931-a5d6-4ec8-9f2b-47be44a5d33a to disappear
Jul 18 19:50:55.545: INFO: Pod downward-api-e8b45931-a5d6-4ec8-9f2b-47be44a5d33a no longer exists
[AfterEach] [sig-node] Downward API
  /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:186
... skipping 4 lines ...
• [SLOW TEST:18.173 seconds]
[sig-node] Downward API
/workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/common/node/framework.go:23
  should provide pod UID as env vars [NodeConformance] [Conformance]
  /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:630
------------------------------
{"msg":"PASSED [sig-node] Downward API should provide pod UID as env vars [NodeConformance] [Conformance]","total":-1,"completed":2,"skipped":18,"failed":0}

SSSSSS
------------------------------
[BeforeEach] [sig-storage] Volumes
  /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:185
STEP: Creating a kubernetes client
... skipping 187 lines ...
• [SLOW TEST:16.304 seconds]
[sig-api-machinery] AdmissionWebhook [Privileged:ClusterAdmin]
/workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/apimachinery/framework.go:23
  should include webhook resources in discovery documents [Conformance]
  /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:630
------------------------------
{"msg":"PASSED [sig-api-machinery] AdmissionWebhook [Privileged:ClusterAdmin] should include webhook resources in discovery documents [Conformance]","total":-1,"completed":2,"skipped":7,"failed":0}
[BeforeEach] [Testpattern: Pre-provisioned PV (default fs)] volumes
  /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/storage/framework/testsuite.go:51
Jul 18 19:51:00.772: INFO: Driver emptydir doesn't support PreprovisionedPV -- skipping
[AfterEach] [Testpattern: Pre-provisioned PV (default fs)] volumes
  /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:186

... skipping 94 lines ...
Jul 18 19:50:51.609: INFO: PersistentVolumeClaim pvc-s4r4r found but phase is Pending instead of Bound.
Jul 18 19:50:53.640: INFO: PersistentVolumeClaim pvc-s4r4r found and phase=Bound (6.131006457s)
Jul 18 19:50:53.640: INFO: Waiting up to 3m0s for PersistentVolume local-vnj92 to have phase Bound
Jul 18 19:50:53.671: INFO: PersistentVolume local-vnj92 found and phase=Bound (30.683466ms)
STEP: Creating pod exec-volume-test-preprovisionedpv-bwrd
STEP: Creating a pod to test exec-volume-test
Jul 18 19:50:53.765: INFO: Waiting up to 5m0s for pod "exec-volume-test-preprovisionedpv-bwrd" in namespace "volume-8483" to be "Succeeded or Failed"
Jul 18 19:50:53.796: INFO: Pod "exec-volume-test-preprovisionedpv-bwrd": Phase="Pending", Reason="", readiness=false. Elapsed: 30.863913ms
Jul 18 19:50:55.828: INFO: Pod "exec-volume-test-preprovisionedpv-bwrd": Phase="Pending", Reason="", readiness=false. Elapsed: 2.063388739s
Jul 18 19:50:57.860: INFO: Pod "exec-volume-test-preprovisionedpv-bwrd": Phase="Pending", Reason="", readiness=false. Elapsed: 4.095416956s
Jul 18 19:50:59.891: INFO: Pod "exec-volume-test-preprovisionedpv-bwrd": Phase="Pending", Reason="", readiness=false. Elapsed: 6.126190442s
Jul 18 19:51:01.923: INFO: Pod "exec-volume-test-preprovisionedpv-bwrd": Phase="Pending", Reason="", readiness=false. Elapsed: 8.158544761s
Jul 18 19:51:03.954: INFO: Pod "exec-volume-test-preprovisionedpv-bwrd": Phase="Pending", Reason="", readiness=false. Elapsed: 10.189028629s
Jul 18 19:51:05.986: INFO: Pod "exec-volume-test-preprovisionedpv-bwrd": Phase="Pending", Reason="", readiness=false. Elapsed: 12.221453889s
Jul 18 19:51:08.262: INFO: Pod "exec-volume-test-preprovisionedpv-bwrd": Phase="Succeeded", Reason="", readiness=false. Elapsed: 14.497593829s
STEP: Saw pod success
Jul 18 19:51:08.262: INFO: Pod "exec-volume-test-preprovisionedpv-bwrd" satisfied condition "Succeeded or Failed"
Jul 18 19:51:08.293: INFO: Trying to get logs from node 10.116.0.7 pod exec-volume-test-preprovisionedpv-bwrd container exec-container-preprovisionedpv-bwrd: <nil>
STEP: delete the pod
Jul 18 19:51:08.374: INFO: Waiting for pod exec-volume-test-preprovisionedpv-bwrd to disappear
Jul 18 19:51:08.406: INFO: Pod exec-volume-test-preprovisionedpv-bwrd no longer exists
STEP: Deleting pod exec-volume-test-preprovisionedpv-bwrd
Jul 18 19:51:08.406: INFO: Deleting pod "exec-volume-test-preprovisionedpv-bwrd" in namespace "volume-8483"
... skipping 22 lines ...
  /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/storage/in_tree_volumes.go:58
    [Testpattern: Pre-provisioned PV (default fs)] volumes
    /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/storage/framework/testsuite.go:50
      should allow exec of files on the volume
      /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/storage/testsuites/volumes.go:196
------------------------------
{"msg":"PASSED [sig-storage] In-tree Volumes [Driver: local][LocalVolumeType: block] [Testpattern: Pre-provisioned PV (default fs)] volumes should allow exec of files on the volume","total":-1,"completed":1,"skipped":5,"failed":0}

S
------------------------------
[BeforeEach] [Testpattern: Dynamic PV (default fs)] provisioning
  /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/storage/framework/testsuite.go:51
Jul 18 19:51:09.494: INFO: Driver local doesn't support DynamicPV -- skipping
... skipping 81 lines ...
  /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/storage/persistent_volumes-local.go:192
    Two pods mounting a local volume one after the other
    /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/storage/persistent_volumes-local.go:254
      should be able to write from pod1 and read from pod2
      /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/storage/persistent_volumes-local.go:255
------------------------------
{"msg":"PASSED [sig-storage] PersistentVolumes-local  [Volume type: dir-link-bindmounted] Two pods mounting a local volume one after the other should be able to write from pod1 and read from pod2","total":-1,"completed":1,"skipped":7,"failed":0}

SSSSSS
------------------------------
[BeforeEach] [Testpattern: Dynamic PV (delayed binding)] topology
  /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/storage/framework/testsuite.go:51
Jul 18 19:51:14.736: INFO: Only supported for node OS distro [gci ubuntu custom] (not debian)
... skipping 5 lines ...
[sig-storage] In-tree Volumes
/workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/storage/utils/framework.go:23
  [Driver: gluster]
  /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/storage/in_tree_volumes.go:58
    [Testpattern: Dynamic PV (delayed binding)] topology
    /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/storage/framework/testsuite.go:50
      should fail to schedule a pod which has topologies that conflict with AllowedTopologies [BeforeEach]
      /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/storage/testsuites/topology.go:192

      Only supported for node OS distro [gci ubuntu custom] (not debian)

      /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/storage/drivers/in_tree.go:263
------------------------------
... skipping 19 lines ...
• [SLOW TEST:26.242 seconds]
[sig-api-machinery] CustomResourcePublishOpenAPI [Privileged:ClusterAdmin]
/workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/apimachinery/framework.go:23
  removes definition from spec when one version gets changed to not be served [Conformance]
  /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:630
------------------------------
{"msg":"PASSED [sig-api-machinery] CustomResourcePublishOpenAPI [Privileged:ClusterAdmin] removes definition from spec when one version gets changed to not be served [Conformance]","total":-1,"completed":2,"skipped":5,"failed":0}

SS
------------------------------
[BeforeEach] [Testpattern: Inline-volume (default fs)] subPath
  /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/storage/framework/testsuite.go:51
Jul 18 19:51:14.829: INFO: Only supported for providers [gce gke] (not skeleton)
... skipping 41 lines ...
Jul 18 19:50:49.778: INFO: Unable to read wheezy_udp@dns-test-service-2.dns-3211.svc.cluster.local from pod dns-3211/dns-test-2eeff506-5444-4d34-a2dd-f276a78c1e05: the server could not find the requested resource (get pods dns-test-2eeff506-5444-4d34-a2dd-f276a78c1e05)
Jul 18 19:50:49.810: INFO: Unable to read wheezy_tcp@dns-test-service-2.dns-3211.svc.cluster.local from pod dns-3211/dns-test-2eeff506-5444-4d34-a2dd-f276a78c1e05: the server could not find the requested resource (get pods dns-test-2eeff506-5444-4d34-a2dd-f276a78c1e05)
Jul 18 19:50:49.938: INFO: Unable to read jessie_udp@dns-querier-2.dns-test-service-2.dns-3211.svc.cluster.local from pod dns-3211/dns-test-2eeff506-5444-4d34-a2dd-f276a78c1e05: the server could not find the requested resource (get pods dns-test-2eeff506-5444-4d34-a2dd-f276a78c1e05)
Jul 18 19:50:49.969: INFO: Unable to read jessie_tcp@dns-querier-2.dns-test-service-2.dns-3211.svc.cluster.local from pod dns-3211/dns-test-2eeff506-5444-4d34-a2dd-f276a78c1e05: the server could not find the requested resource (get pods dns-test-2eeff506-5444-4d34-a2dd-f276a78c1e05)
Jul 18 19:50:50.001: INFO: Unable to read jessie_udp@dns-test-service-2.dns-3211.svc.cluster.local from pod dns-3211/dns-test-2eeff506-5444-4d34-a2dd-f276a78c1e05: the server could not find the requested resource (get pods dns-test-2eeff506-5444-4d34-a2dd-f276a78c1e05)
Jul 18 19:50:50.032: INFO: Unable to read jessie_tcp@dns-test-service-2.dns-3211.svc.cluster.local from pod dns-3211/dns-test-2eeff506-5444-4d34-a2dd-f276a78c1e05: the server could not find the requested resource (get pods dns-test-2eeff506-5444-4d34-a2dd-f276a78c1e05)
Jul 18 19:50:50.095: INFO: Lookups using dns-3211/dns-test-2eeff506-5444-4d34-a2dd-f276a78c1e05 failed for: [wheezy_udp@dns-querier-2.dns-test-service-2.dns-3211.svc.cluster.local wheezy_tcp@dns-querier-2.dns-test-service-2.dns-3211.svc.cluster.local wheezy_udp@dns-test-service-2.dns-3211.svc.cluster.local wheezy_tcp@dns-test-service-2.dns-3211.svc.cluster.local jessie_udp@dns-querier-2.dns-test-service-2.dns-3211.svc.cluster.local jessie_tcp@dns-querier-2.dns-test-service-2.dns-3211.svc.cluster.local jessie_udp@dns-test-service-2.dns-3211.svc.cluster.local jessie_tcp@dns-test-service-2.dns-3211.svc.cluster.local]

Jul 18 19:50:55.126: INFO: Unable to read wheezy_udp@dns-querier-2.dns-test-service-2.dns-3211.svc.cluster.local from pod dns-3211/dns-test-2eeff506-5444-4d34-a2dd-f276a78c1e05: the server could not find the requested resource (get pods dns-test-2eeff506-5444-4d34-a2dd-f276a78c1e05)
Jul 18 19:50:55.158: INFO: Unable to read wheezy_tcp@dns-querier-2.dns-test-service-2.dns-3211.svc.cluster.local from pod dns-3211/dns-test-2eeff506-5444-4d34-a2dd-f276a78c1e05: the server could not find the requested resource (get pods dns-test-2eeff506-5444-4d34-a2dd-f276a78c1e05)
Jul 18 19:50:55.188: INFO: Unable to read wheezy_udp@dns-test-service-2.dns-3211.svc.cluster.local from pod dns-3211/dns-test-2eeff506-5444-4d34-a2dd-f276a78c1e05: the server could not find the requested resource (get pods dns-test-2eeff506-5444-4d34-a2dd-f276a78c1e05)
Jul 18 19:50:55.220: INFO: Unable to read wheezy_tcp@dns-test-service-2.dns-3211.svc.cluster.local from pod dns-3211/dns-test-2eeff506-5444-4d34-a2dd-f276a78c1e05: the server could not find the requested resource (get pods dns-test-2eeff506-5444-4d34-a2dd-f276a78c1e05)
Jul 18 19:50:55.314: INFO: Unable to read jessie_udp@dns-querier-2.dns-test-service-2.dns-3211.svc.cluster.local from pod dns-3211/dns-test-2eeff506-5444-4d34-a2dd-f276a78c1e05: the server could not find the requested resource (get pods dns-test-2eeff506-5444-4d34-a2dd-f276a78c1e05)
Jul 18 19:50:55.345: INFO: Unable to read jessie_tcp@dns-querier-2.dns-test-service-2.dns-3211.svc.cluster.local from pod dns-3211/dns-test-2eeff506-5444-4d34-a2dd-f276a78c1e05: the server could not find the requested resource (get pods dns-test-2eeff506-5444-4d34-a2dd-f276a78c1e05)
Jul 18 19:50:55.377: INFO: Unable to read jessie_udp@dns-test-service-2.dns-3211.svc.cluster.local from pod dns-3211/dns-test-2eeff506-5444-4d34-a2dd-f276a78c1e05: the server could not find the requested resource (get pods dns-test-2eeff506-5444-4d34-a2dd-f276a78c1e05)
Jul 18 19:50:55.408: INFO: Unable to read jessie_tcp@dns-test-service-2.dns-3211.svc.cluster.local from pod dns-3211/dns-test-2eeff506-5444-4d34-a2dd-f276a78c1e05: the server could not find the requested resource (get pods dns-test-2eeff506-5444-4d34-a2dd-f276a78c1e05)
Jul 18 19:50:55.473: INFO: Lookups using dns-3211/dns-test-2eeff506-5444-4d34-a2dd-f276a78c1e05 failed for: [wheezy_udp@dns-querier-2.dns-test-service-2.dns-3211.svc.cluster.local wheezy_tcp@dns-querier-2.dns-test-service-2.dns-3211.svc.cluster.local wheezy_udp@dns-test-service-2.dns-3211.svc.cluster.local wheezy_tcp@dns-test-service-2.dns-3211.svc.cluster.local jessie_udp@dns-querier-2.dns-test-service-2.dns-3211.svc.cluster.local jessie_tcp@dns-querier-2.dns-test-service-2.dns-3211.svc.cluster.local jessie_udp@dns-test-service-2.dns-3211.svc.cluster.local jessie_tcp@dns-test-service-2.dns-3211.svc.cluster.local]

Jul 18 19:51:00.128: INFO: Unable to read wheezy_udp@dns-querier-2.dns-test-service-2.dns-3211.svc.cluster.local from pod dns-3211/dns-test-2eeff506-5444-4d34-a2dd-f276a78c1e05: the server could not find the requested resource (get pods dns-test-2eeff506-5444-4d34-a2dd-f276a78c1e05)
Jul 18 19:51:00.160: INFO: Unable to read wheezy_tcp@dns-querier-2.dns-test-service-2.dns-3211.svc.cluster.local from pod dns-3211/dns-test-2eeff506-5444-4d34-a2dd-f276a78c1e05: the server could not find the requested resource (get pods dns-test-2eeff506-5444-4d34-a2dd-f276a78c1e05)
Jul 18 19:51:00.192: INFO: Unable to read wheezy_udp@dns-test-service-2.dns-3211.svc.cluster.local from pod dns-3211/dns-test-2eeff506-5444-4d34-a2dd-f276a78c1e05: the server could not find the requested resource (get pods dns-test-2eeff506-5444-4d34-a2dd-f276a78c1e05)
Jul 18 19:51:00.224: INFO: Unable to read wheezy_tcp@dns-test-service-2.dns-3211.svc.cluster.local from pod dns-3211/dns-test-2eeff506-5444-4d34-a2dd-f276a78c1e05: the server could not find the requested resource (get pods dns-test-2eeff506-5444-4d34-a2dd-f276a78c1e05)
Jul 18 19:51:00.320: INFO: Unable to read jessie_udp@dns-querier-2.dns-test-service-2.dns-3211.svc.cluster.local from pod dns-3211/dns-test-2eeff506-5444-4d34-a2dd-f276a78c1e05: the server could not find the requested resource (get pods dns-test-2eeff506-5444-4d34-a2dd-f276a78c1e05)
Jul 18 19:51:00.354: INFO: Unable to read jessie_tcp@dns-querier-2.dns-test-service-2.dns-3211.svc.cluster.local from pod dns-3211/dns-test-2eeff506-5444-4d34-a2dd-f276a78c1e05: the server could not find the requested resource (get pods dns-test-2eeff506-5444-4d34-a2dd-f276a78c1e05)
Jul 18 19:51:00.388: INFO: Unable to read jessie_udp@dns-test-service-2.dns-3211.svc.cluster.local from pod dns-3211/dns-test-2eeff506-5444-4d34-a2dd-f276a78c1e05: the server could not find the requested resource (get pods dns-test-2eeff506-5444-4d34-a2dd-f276a78c1e05)
Jul 18 19:51:00.420: INFO: Unable to read jessie_tcp@dns-test-service-2.dns-3211.svc.cluster.local from pod dns-3211/dns-test-2eeff506-5444-4d34-a2dd-f276a78c1e05: the server could not find the requested resource (get pods dns-test-2eeff506-5444-4d34-a2dd-f276a78c1e05)
Jul 18 19:51:00.483: INFO: Lookups using dns-3211/dns-test-2eeff506-5444-4d34-a2dd-f276a78c1e05 failed for: [wheezy_udp@dns-querier-2.dns-test-service-2.dns-3211.svc.cluster.local wheezy_tcp@dns-querier-2.dns-test-service-2.dns-3211.svc.cluster.local wheezy_udp@dns-test-service-2.dns-3211.svc.cluster.local wheezy_tcp@dns-test-service-2.dns-3211.svc.cluster.local jessie_udp@dns-querier-2.dns-test-service-2.dns-3211.svc.cluster.local jessie_tcp@dns-querier-2.dns-test-service-2.dns-3211.svc.cluster.local jessie_udp@dns-test-service-2.dns-3211.svc.cluster.local jessie_tcp@dns-test-service-2.dns-3211.svc.cluster.local]

Jul 18 19:51:05.128: INFO: Unable to read wheezy_udp@dns-querier-2.dns-test-service-2.dns-3211.svc.cluster.local from pod dns-3211/dns-test-2eeff506-5444-4d34-a2dd-f276a78c1e05: the server could not find the requested resource (get pods dns-test-2eeff506-5444-4d34-a2dd-f276a78c1e05)
Jul 18 19:51:05.160: INFO: Unable to read wheezy_tcp@dns-querier-2.dns-test-service-2.dns-3211.svc.cluster.local from pod dns-3211/dns-test-2eeff506-5444-4d34-a2dd-f276a78c1e05: the server could not find the requested resource (get pods dns-test-2eeff506-5444-4d34-a2dd-f276a78c1e05)
Jul 18 19:51:05.192: INFO: Unable to read wheezy_udp@dns-test-service-2.dns-3211.svc.cluster.local from pod dns-3211/dns-test-2eeff506-5444-4d34-a2dd-f276a78c1e05: the server could not find the requested resource (get pods dns-test-2eeff506-5444-4d34-a2dd-f276a78c1e05)
Jul 18 19:51:05.224: INFO: Unable to read wheezy_tcp@dns-test-service-2.dns-3211.svc.cluster.local from pod dns-3211/dns-test-2eeff506-5444-4d34-a2dd-f276a78c1e05: the server could not find the requested resource (get pods dns-test-2eeff506-5444-4d34-a2dd-f276a78c1e05)
Jul 18 19:51:05.321: INFO: Unable to read jessie_udp@dns-querier-2.dns-test-service-2.dns-3211.svc.cluster.local from pod dns-3211/dns-test-2eeff506-5444-4d34-a2dd-f276a78c1e05: the server could not find the requested resource (get pods dns-test-2eeff506-5444-4d34-a2dd-f276a78c1e05)
Jul 18 19:51:05.353: INFO: Unable to read jessie_tcp@dns-querier-2.dns-test-service-2.dns-3211.svc.cluster.local from pod dns-3211/dns-test-2eeff506-5444-4d34-a2dd-f276a78c1e05: the server could not find the requested resource (get pods dns-test-2eeff506-5444-4d34-a2dd-f276a78c1e05)
Jul 18 19:51:05.386: INFO: Unable to read jessie_udp@dns-test-service-2.dns-3211.svc.cluster.local from pod dns-3211/dns-test-2eeff506-5444-4d34-a2dd-f276a78c1e05: the server could not find the requested resource (get pods dns-test-2eeff506-5444-4d34-a2dd-f276a78c1e05)
Jul 18 19:51:05.418: INFO: Unable to read jessie_tcp@dns-test-service-2.dns-3211.svc.cluster.local from pod dns-3211/dns-test-2eeff506-5444-4d34-a2dd-f276a78c1e05: the server could not find the requested resource (get pods dns-test-2eeff506-5444-4d34-a2dd-f276a78c1e05)
Jul 18 19:51:05.481: INFO: Lookups using dns-3211/dns-test-2eeff506-5444-4d34-a2dd-f276a78c1e05 failed for: [wheezy_udp@dns-querier-2.dns-test-service-2.dns-3211.svc.cluster.local wheezy_tcp@dns-querier-2.dns-test-service-2.dns-3211.svc.cluster.local wheezy_udp@dns-test-service-2.dns-3211.svc.cluster.local wheezy_tcp@dns-test-service-2.dns-3211.svc.cluster.local jessie_udp@dns-querier-2.dns-test-service-2.dns-3211.svc.cluster.local jessie_tcp@dns-querier-2.dns-test-service-2.dns-3211.svc.cluster.local jessie_udp@dns-test-service-2.dns-3211.svc.cluster.local jessie_tcp@dns-test-service-2.dns-3211.svc.cluster.local]

Jul 18 19:51:10.393: INFO: Unable to read jessie_udp@dns-test-service-2.dns-3211.svc.cluster.local from pod dns-3211/dns-test-2eeff506-5444-4d34-a2dd-f276a78c1e05: the server could not find the requested resource (get pods dns-test-2eeff506-5444-4d34-a2dd-f276a78c1e05)
Jul 18 19:51:10.434: INFO: Unable to read jessie_tcp@dns-test-service-2.dns-3211.svc.cluster.local from pod dns-3211/dns-test-2eeff506-5444-4d34-a2dd-f276a78c1e05: the server could not find the requested resource (get pods dns-test-2eeff506-5444-4d34-a2dd-f276a78c1e05)
Jul 18 19:51:10.499: INFO: Lookups using dns-3211/dns-test-2eeff506-5444-4d34-a2dd-f276a78c1e05 failed for: [jessie_udp@dns-test-service-2.dns-3211.svc.cluster.local jessie_tcp@dns-test-service-2.dns-3211.svc.cluster.local]

Jul 18 19:51:15.594: INFO: DNS probes using dns-3211/dns-test-2eeff506-5444-4d34-a2dd-f276a78c1e05 succeeded

STEP: deleting the pod
STEP: deleting the test headless service
[AfterEach] [sig-network] DNS
... skipping 5 lines ...
• [SLOW TEST:40.497 seconds]
[sig-network] DNS
/workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/network/common/framework.go:23
  should provide DNS for pods for Subdomain [Conformance]
  /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:630
------------------------------
{"msg":"PASSED [sig-network] DNS should provide DNS for pods for Subdomain [Conformance]","total":-1,"completed":1,"skipped":1,"failed":0}

SSSSSS
------------------------------
[BeforeEach] [sig-apps] Job
  /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:185
STEP: Creating a kubernetes client
Jul 18 19:50:50.071: INFO: >>> kubeConfig: /root/.kube/config
STEP: Building a namespace api object, basename job
STEP: Waiting for a default service account to be provisioned in namespace
[It] should run a job to completion when tasks sometimes fail and are locally restarted [Conformance]
  /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:630
STEP: Creating a job
STEP: Ensuring job reaches completions
[AfterEach] [sig-apps] Job
  /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:186
Jul 18 19:51:16.292: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready
STEP: Destroying namespace "job-3079" for this suite.


• [SLOW TEST:26.321 seconds]
[sig-apps] Job
/workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/apps/framework.go:23
  should run a job to completion when tasks sometimes fail and are locally restarted [Conformance]
  /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:630
------------------------------
{"msg":"PASSED [sig-apps] Job should run a job to completion when tasks sometimes fail and are locally restarted [Conformance]","total":-1,"completed":4,"skipped":18,"failed":0}

SS
------------------------------
[BeforeEach] [Testpattern: Inline-volume (default fs)] subPath
  /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/storage/framework/testsuite.go:51
Jul 18 19:51:16.411: INFO: Only supported for providers [aws] (not skeleton)
... skipping 293 lines ...
      Only supported for providers [openstack] (not skeleton)

      /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/storage/drivers/in_tree.go:1092
------------------------------
SSSS
------------------------------
{"msg":"PASSED [sig-cli] Kubectl client Kubectl label should update the label on a resource  [Conformance]","total":-1,"completed":1,"skipped":1,"failed":0}
[BeforeEach] [sig-node] Container Lifecycle Hook
  /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:185
STEP: Creating a kubernetes client
Jul 18 19:50:50.234: INFO: >>> kubeConfig: /root/.kube/config
STEP: Building a namespace api object, basename container-lifecycle-hook
STEP: Waiting for a default service account to be provisioned in namespace
... skipping 40 lines ...
/workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/common/node/framework.go:23
  when create a pod with lifecycle hook
  /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/common/node/lifecycle_hook.go:43
    should execute prestop http hook properly [NodeConformance] [Conformance]
    /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:630
------------------------------
{"msg":"PASSED [sig-node] Container Lifecycle Hook when create a pod with lifecycle hook should execute prestop http hook properly [NodeConformance] [Conformance]","total":-1,"completed":2,"skipped":1,"failed":0}
[BeforeEach] [Testpattern: Dynamic PV (default fs)] subPath
  /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/storage/framework/testsuite.go:51
Jul 18 19:51:20.901: INFO: Only supported for providers [gce gke] (not skeleton)
[AfterEach] [Testpattern: Dynamic PV (default fs)] subPath
  /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:186

... skipping 25 lines ...
[sig-storage] In-tree Volumes
/workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/storage/utils/framework.go:23
  [Driver: local][LocalVolumeType: blockfs]
  /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/storage/in_tree_volumes.go:58
    [Testpattern: Dynamic PV (immediate binding)] topology
    /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/storage/framework/testsuite.go:50
      should fail to schedule a pod which has topologies that conflict with AllowedTopologies [BeforeEach]
      /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/storage/testsuites/topology.go:192

      Driver local doesn't support DynamicPV -- skipping

      /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/storage/framework/testsuite.go:116
------------------------------
... skipping 19 lines ...
      Driver local doesn't support DynamicPV -- skipping

      /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/storage/framework/testsuite.go:116
------------------------------
SSSSS
------------------------------
{"msg":"PASSED [sig-storage] In-tree Volumes [Driver: emptydir] [Testpattern: Inline-volume (default fs)] subPath should support existing single file [LinuxOnly]","total":-1,"completed":2,"skipped":35,"failed":0}
[BeforeEach] [Testpattern: Pre-provisioned PV (default fs)] subPath
  /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/storage/framework/testsuite.go:51
[BeforeEach] [Testpattern: Pre-provisioned PV (default fs)] subPath
  /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:185
STEP: Creating a kubernetes client
Jul 18 19:50:48.317: INFO: >>> kubeConfig: /root/.kube/config
... skipping 13 lines ...
Jul 18 19:51:06.995: INFO: PersistentVolumeClaim pvc-5nl4g found but phase is Pending instead of Bound.
Jul 18 19:51:09.026: INFO: PersistentVolumeClaim pvc-5nl4g found and phase=Bound (4.093746315s)
Jul 18 19:51:09.026: INFO: Waiting up to 3m0s for PersistentVolume local-hz4rm to have phase Bound
Jul 18 19:51:09.061: INFO: PersistentVolume local-hz4rm found and phase=Bound (34.910549ms)
STEP: Creating pod pod-subpath-test-preprovisionedpv-ck5r
STEP: Creating a pod to test subpath
Jul 18 19:51:09.159: INFO: Waiting up to 5m0s for pod "pod-subpath-test-preprovisionedpv-ck5r" in namespace "provisioning-307" to be "Succeeded or Failed"
Jul 18 19:51:09.191: INFO: Pod "pod-subpath-test-preprovisionedpv-ck5r": Phase="Pending", Reason="", readiness=false. Elapsed: 31.288338ms
Jul 18 19:51:11.224: INFO: Pod "pod-subpath-test-preprovisionedpv-ck5r": Phase="Pending", Reason="", readiness=false. Elapsed: 2.064347843s
Jul 18 19:51:13.257: INFO: Pod "pod-subpath-test-preprovisionedpv-ck5r": Phase="Pending", Reason="", readiness=false. Elapsed: 4.097958422s
Jul 18 19:51:15.290: INFO: Pod "pod-subpath-test-preprovisionedpv-ck5r": Phase="Pending", Reason="", readiness=false. Elapsed: 6.130410494s
Jul 18 19:51:17.323: INFO: Pod "pod-subpath-test-preprovisionedpv-ck5r": Phase="Pending", Reason="", readiness=false. Elapsed: 8.163361257s
Jul 18 19:51:19.355: INFO: Pod "pod-subpath-test-preprovisionedpv-ck5r": Phase="Pending", Reason="", readiness=false. Elapsed: 10.195576237s
Jul 18 19:51:21.386: INFO: Pod "pod-subpath-test-preprovisionedpv-ck5r": Phase="Succeeded", Reason="", readiness=false. Elapsed: 12.226951503s
STEP: Saw pod success
Jul 18 19:51:21.386: INFO: Pod "pod-subpath-test-preprovisionedpv-ck5r" satisfied condition "Succeeded or Failed"
Jul 18 19:51:21.417: INFO: Trying to get logs from node 10.116.0.5 pod pod-subpath-test-preprovisionedpv-ck5r container test-container-subpath-preprovisionedpv-ck5r: <nil>
STEP: delete the pod
Jul 18 19:51:21.492: INFO: Waiting for pod pod-subpath-test-preprovisionedpv-ck5r to disappear
Jul 18 19:51:21.522: INFO: Pod pod-subpath-test-preprovisionedpv-ck5r no longer exists
STEP: Deleting pod pod-subpath-test-preprovisionedpv-ck5r
Jul 18 19:51:21.522: INFO: Deleting pod "pod-subpath-test-preprovisionedpv-ck5r" in namespace "provisioning-307"
... skipping 19 lines ...
  /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/storage/in_tree_volumes.go:58
    [Testpattern: Pre-provisioned PV (default fs)] subPath
    /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/storage/framework/testsuite.go:50
      should support existing single file [LinuxOnly]
      /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/storage/testsuites/subpath.go:219
------------------------------
{"msg":"PASSED [sig-storage] In-tree Volumes [Driver: local][LocalVolumeType: dir] [Testpattern: Pre-provisioned PV (default fs)] subPath should support existing single file [LinuxOnly]","total":-1,"completed":3,"skipped":35,"failed":0}

SSS
------------------------------
[BeforeEach] [Testpattern: Pre-provisioned PV (block volmode)] volumeMode
  /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/storage/framework/testsuite.go:51
Jul 18 19:51:22.102: INFO: Only supported for providers [gce gke] (not skeleton)
... skipping 70 lines ...
• [SLOW TEST:32.698 seconds]
[sig-apps] Deployment
/workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/apps/framework.go:23
  deployment should delete old replica sets [Conformance]
  /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:630
------------------------------
{"msg":"PASSED [sig-apps] Deployment deployment should delete old replica sets [Conformance]","total":-1,"completed":2,"skipped":49,"failed":0}

SSSSSSSS
------------------------------
[BeforeEach] [Testpattern: Dynamic PV (block volmode)(allowExpansion)] volume-expand
  /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/storage/framework/testsuite.go:51
Jul 18 19:51:25.170: INFO: Driver emptydir doesn't support DynamicPV -- skipping
... skipping 200 lines ...
  /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:186
Jul 18 19:51:26.442: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready
STEP: Destroying namespace "emptydir-698" for this suite.

•
------------------------------
{"msg":"PASSED [sig-storage] EmptyDir volumes pod should support memory backed volumes of specified size","total":-1,"completed":1,"skipped":10,"failed":0}
[BeforeEach] [Testpattern: Inline-volume (ext4)] volumes
  /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/storage/framework/testsuite.go:51
Jul 18 19:51:26.519: INFO: Driver hostPath doesn't support ext4 -- skipping
[AfterEach] [Testpattern: Inline-volume (ext4)] volumes
  /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:186

... skipping 35 lines ...
• [SLOW TEST:10.855 seconds]
[sig-auth] Certificates API [Privileged:ClusterAdmin]
/workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/auth/framework.go:23
  should support building a client with a CSR
  /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/auth/certificates.go:55
------------------------------
{"msg":"PASSED [sig-auth] Certificates API [Privileged:ClusterAdmin] should support building a client with a CSR","total":-1,"completed":5,"skipped":35,"failed":0}

S
------------------------------
[BeforeEach] [Testpattern: Dynamic PV (default fs)] subPath
  /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/storage/framework/testsuite.go:51
Jul 18 19:51:27.367: INFO: Only supported for providers [openstack] (not skeleton)
... skipping 147 lines ...
Jul 18 19:51:06.252: INFO: PersistentVolumeClaim pvc-z4kvd found but phase is Pending instead of Bound.
Jul 18 19:51:08.284: INFO: PersistentVolumeClaim pvc-z4kvd found and phase=Bound (16.296218207s)
Jul 18 19:51:08.284: INFO: Waiting up to 3m0s for PersistentVolume local-gc767 to have phase Bound
Jul 18 19:51:08.318: INFO: PersistentVolume local-gc767 found and phase=Bound (33.252822ms)
STEP: Creating pod pod-subpath-test-preprovisionedpv-d89h
STEP: Creating a pod to test subpath
Jul 18 19:51:08.423: INFO: Waiting up to 5m0s for pod "pod-subpath-test-preprovisionedpv-d89h" in namespace "provisioning-6448" to be "Succeeded or Failed"
Jul 18 19:51:08.456: INFO: Pod "pod-subpath-test-preprovisionedpv-d89h": Phase="Pending", Reason="", readiness=false. Elapsed: 32.328428ms
Jul 18 19:51:10.489: INFO: Pod "pod-subpath-test-preprovisionedpv-d89h": Phase="Pending", Reason="", readiness=false. Elapsed: 2.065003953s
Jul 18 19:51:12.521: INFO: Pod "pod-subpath-test-preprovisionedpv-d89h": Phase="Pending", Reason="", readiness=false. Elapsed: 4.097563614s
Jul 18 19:51:14.554: INFO: Pod "pod-subpath-test-preprovisionedpv-d89h": Phase="Succeeded", Reason="", readiness=false. Elapsed: 6.130426552s
STEP: Saw pod success
Jul 18 19:51:14.554: INFO: Pod "pod-subpath-test-preprovisionedpv-d89h" satisfied condition "Succeeded or Failed"
Jul 18 19:51:14.585: INFO: Trying to get logs from node 10.116.0.7 pod pod-subpath-test-preprovisionedpv-d89h container test-container-subpath-preprovisionedpv-d89h: <nil>
STEP: delete the pod
Jul 18 19:51:14.669: INFO: Waiting for pod pod-subpath-test-preprovisionedpv-d89h to disappear
Jul 18 19:51:14.705: INFO: Pod pod-subpath-test-preprovisionedpv-d89h no longer exists
STEP: Deleting pod pod-subpath-test-preprovisionedpv-d89h
Jul 18 19:51:14.705: INFO: Deleting pod "pod-subpath-test-preprovisionedpv-d89h" in namespace "provisioning-6448"
STEP: Creating pod pod-subpath-test-preprovisionedpv-d89h
STEP: Creating a pod to test subpath
Jul 18 19:51:14.771: INFO: Waiting up to 5m0s for pod "pod-subpath-test-preprovisionedpv-d89h" in namespace "provisioning-6448" to be "Succeeded or Failed"
Jul 18 19:51:14.804: INFO: Pod "pod-subpath-test-preprovisionedpv-d89h": Phase="Pending", Reason="", readiness=false. Elapsed: 33.382971ms
Jul 18 19:51:16.837: INFO: Pod "pod-subpath-test-preprovisionedpv-d89h": Phase="Pending", Reason="", readiness=false. Elapsed: 2.066249944s
Jul 18 19:51:18.870: INFO: Pod "pod-subpath-test-preprovisionedpv-d89h": Phase="Pending", Reason="", readiness=false. Elapsed: 4.099433373s
Jul 18 19:51:20.903: INFO: Pod "pod-subpath-test-preprovisionedpv-d89h": Phase="Pending", Reason="", readiness=false. Elapsed: 6.132082605s
Jul 18 19:51:22.939: INFO: Pod "pod-subpath-test-preprovisionedpv-d89h": Phase="Pending", Reason="", readiness=false. Elapsed: 8.168087154s
Jul 18 19:51:24.976: INFO: Pod "pod-subpath-test-preprovisionedpv-d89h": Phase="Pending", Reason="", readiness=false. Elapsed: 10.20501424s
Jul 18 19:51:27.008: INFO: Pod "pod-subpath-test-preprovisionedpv-d89h": Phase="Pending", Reason="", readiness=false. Elapsed: 12.237529179s
Jul 18 19:51:29.044: INFO: Pod "pod-subpath-test-preprovisionedpv-d89h": Phase="Pending", Reason="", readiness=false. Elapsed: 14.273531048s
Jul 18 19:51:31.076: INFO: Pod "pod-subpath-test-preprovisionedpv-d89h": Phase="Succeeded", Reason="", readiness=false. Elapsed: 16.305172322s
STEP: Saw pod success
Jul 18 19:51:31.076: INFO: Pod "pod-subpath-test-preprovisionedpv-d89h" satisfied condition "Succeeded or Failed"
Jul 18 19:51:31.109: INFO: Trying to get logs from node 10.116.0.7 pod pod-subpath-test-preprovisionedpv-d89h container test-container-subpath-preprovisionedpv-d89h: <nil>
STEP: delete the pod
Jul 18 19:51:31.189: INFO: Waiting for pod pod-subpath-test-preprovisionedpv-d89h to disappear
Jul 18 19:51:31.224: INFO: Pod pod-subpath-test-preprovisionedpv-d89h no longer exists
STEP: Deleting pod pod-subpath-test-preprovisionedpv-d89h
Jul 18 19:51:31.224: INFO: Deleting pod "pod-subpath-test-preprovisionedpv-d89h" in namespace "provisioning-6448"
... skipping 19 lines ...
  /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/storage/in_tree_volumes.go:58
    [Testpattern: Pre-provisioned PV (default fs)] subPath
    /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/storage/framework/testsuite.go:50
      should support existing directories when readOnly specified in the volumeSource
      /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/storage/testsuites/subpath.go:394
------------------------------
{"msg":"PASSED [sig-storage] In-tree Volumes [Driver: local][LocalVolumeType: dir-bindmounted] [Testpattern: Pre-provisioned PV (default fs)] subPath should support existing directories when readOnly specified in the volumeSource","total":-1,"completed":1,"skipped":1,"failed":0}
[BeforeEach] [Testpattern: Pre-provisioned PV (default fs)] subPath
  /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/storage/framework/testsuite.go:51
Jul 18 19:51:31.789: INFO: Driver hostPathSymlink doesn't support PreprovisionedPV -- skipping
[AfterEach] [Testpattern: Pre-provisioned PV (default fs)] subPath
  /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:186

... skipping 55 lines ...
• [SLOW TEST:19.093 seconds]
[sig-api-machinery] AdmissionWebhook [Privileged:ClusterAdmin]
/workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/apimachinery/framework.go:23
  should be able to deny custom resource creation, update and deletion [Conformance]
  /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:630
------------------------------
{"msg":"PASSED [sig-api-machinery] AdmissionWebhook [Privileged:ClusterAdmin] should be able to deny custom resource creation, update and deletion [Conformance]","total":-1,"completed":2,"skipped":14,"failed":0}

S
------------------------------
[BeforeEach] [Testpattern: Dynamic PV (default fs)] fsgroupchangepolicy
  /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/storage/framework/testsuite.go:51
Jul 18 19:51:33.852: INFO: Driver hostPath doesn't support DynamicPV -- skipping
... skipping 55 lines ...
• [SLOW TEST:60.641 seconds]
[sig-api-machinery] Watchers
/workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/apimachinery/framework.go:23
  should observe add, update, and delete watch notifications on configmaps [Conformance]
  /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:630
------------------------------
{"msg":"PASSED [sig-api-machinery] Watchers should observe add, update, and delete watch notifications on configmaps [Conformance]","total":-1,"completed":1,"skipped":1,"failed":0}

SS
------------------------------
[BeforeEach] [sig-network] Services
  /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:185
STEP: Creating a kubernetes client
... skipping 45 lines ...
• [SLOW TEST:27.277 seconds]
[sig-network] Services
/workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/network/common/framework.go:23
  should serve multiport endpoints from pods  [Conformance]
  /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:630
------------------------------
{"msg":"PASSED [sig-network] Services should serve multiport endpoints from pods  [Conformance]","total":-1,"completed":2,"skipped":8,"failed":0}
[BeforeEach] [Testpattern: Pre-provisioned PV (default fs)] subPath
  /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/storage/framework/testsuite.go:51
Jul 18 19:51:36.791: INFO: Only supported for providers [gce gke] (not skeleton)
[AfterEach] [Testpattern: Pre-provisioned PV (default fs)] subPath
  /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:186

... skipping 128 lines ...
• [SLOW TEST:18.026 seconds]
[sig-api-machinery] AdmissionWebhook [Privileged:ClusterAdmin]
/workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/apimachinery/framework.go:23
  should mutate custom resource [Conformance]
  /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:630
------------------------------
{"msg":"PASSED [sig-api-machinery] AdmissionWebhook [Privileged:ClusterAdmin] should mutate custom resource [Conformance]","total":-1,"completed":3,"skipped":68,"failed":0}

SSSSSSSSS
------------------------------
[BeforeEach] [sig-cli] Kubectl client
  /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:185
STEP: Creating a kubernetes client
... skipping 11 lines ...
  /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:186
Jul 18 19:51:43.508: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready
STEP: Destroying namespace "kubectl-2354" for this suite.

•
------------------------------
{"msg":"PASSED [sig-cli] Kubectl client Proxy server should support --unix-socket=/path  [Conformance]","total":-1,"completed":4,"skipped":77,"failed":0}

S
------------------------------
[BeforeEach] [Testpattern: Dynamic PV (immediate binding)] topology
  /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/storage/framework/testsuite.go:51
Jul 18 19:51:43.592: INFO: Driver local doesn't support DynamicPV -- skipping
... skipping 5 lines ...
[sig-storage] In-tree Volumes
/workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/storage/utils/framework.go:23
  [Driver: local][LocalVolumeType: dir-link-bindmounted]
  /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/storage/in_tree_volumes.go:58
    [Testpattern: Dynamic PV (immediate binding)] topology
    /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/storage/framework/testsuite.go:50
      should fail to schedule a pod which has topologies that conflict with AllowedTopologies [BeforeEach]
      /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/storage/testsuites/topology.go:192

      Driver local doesn't support DynamicPV -- skipping

      /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/storage/framework/testsuite.go:116
------------------------------
... skipping 17 lines ...
      Only supported for providers [openstack] (not skeleton)

      /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/storage/drivers/in_tree.go:1092
------------------------------
SSSS
------------------------------
{"msg":"PASSED [sig-storage] In-tree Volumes [Driver: local][LocalVolumeType: tmpfs] [Testpattern: Pre-provisioned PV (filesystem volmode)] volumeMode should not mount / map unused volumes in a pod [LinuxOnly]","total":-1,"completed":1,"skipped":6,"failed":0}
[BeforeEach] [sig-apps] Deployment
  /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:185
STEP: Creating a kubernetes client
Jul 18 19:51:19.008: INFO: >>> kubeConfig: /root/.kube/config
STEP: Building a namespace api object, basename deployment
STEP: Waiting for a default service account to be provisioned in namespace
... skipping 106 lines ...
• [SLOW TEST:25.241 seconds]
[sig-apps] Deployment
/workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/apps/framework.go:23
  deployment should support proportional scaling [Conformance]
  /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:630
------------------------------
{"msg":"PASSED [sig-apps] Deployment deployment should support proportional scaling [Conformance]","total":-1,"completed":2,"skipped":6,"failed":0}

S
------------------------------
[BeforeEach] [sig-storage] CSI mock volume
  /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:185
STEP: Creating a kubernetes client
... skipping 108 lines ...
/workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/storage/utils/framework.go:23
  CSI attach test using mock driver
  /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/storage/csi_mock_volume.go:316
    should require VolumeAttach for drivers with attachment
    /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/storage/csi_mock_volume.go:338
------------------------------
{"msg":"PASSED [sig-storage] CSI mock volume CSI attach test using mock driver should require VolumeAttach for drivers with attachment","total":-1,"completed":1,"skipped":5,"failed":0}

SSS
------------------------------
[BeforeEach] [Testpattern: Inline-volume (ext4)] volumes
  /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/storage/framework/testsuite.go:51
Jul 18 19:51:50.327: INFO: Only supported for providers [vsphere] (not skeleton)
... skipping 23 lines ...
Jul 18 19:51:35.928: INFO: >>> kubeConfig: /root/.kube/config
STEP: Building a namespace api object, basename downward-api
STEP: Waiting for a default service account to be provisioned in namespace
[It] should provide host IP and pod IP as an env var if pod uses host network [LinuxOnly]
  /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/common/node/downwardapi.go:109
STEP: Creating a pod to test downward api env vars
Jul 18 19:51:36.118: INFO: Waiting up to 5m0s for pod "downward-api-357b4e9c-27b4-4c57-875a-2032c7b53b53" in namespace "downward-api-2821" to be "Succeeded or Failed"
Jul 18 19:51:36.148: INFO: Pod "downward-api-357b4e9c-27b4-4c57-875a-2032c7b53b53": Phase="Pending", Reason="", readiness=false. Elapsed: 30.050319ms
Jul 18 19:51:38.178: INFO: Pod "downward-api-357b4e9c-27b4-4c57-875a-2032c7b53b53": Phase="Pending", Reason="", readiness=false. Elapsed: 2.060333179s
Jul 18 19:51:40.209: INFO: Pod "downward-api-357b4e9c-27b4-4c57-875a-2032c7b53b53": Phase="Pending", Reason="", readiness=false. Elapsed: 4.090961148s
Jul 18 19:51:42.242: INFO: Pod "downward-api-357b4e9c-27b4-4c57-875a-2032c7b53b53": Phase="Pending", Reason="", readiness=false. Elapsed: 6.124059008s
Jul 18 19:51:44.273: INFO: Pod "downward-api-357b4e9c-27b4-4c57-875a-2032c7b53b53": Phase="Pending", Reason="", readiness=false. Elapsed: 8.154755237s
Jul 18 19:51:46.305: INFO: Pod "downward-api-357b4e9c-27b4-4c57-875a-2032c7b53b53": Phase="Pending", Reason="", readiness=false. Elapsed: 10.186861432s
Jul 18 19:51:48.336: INFO: Pod "downward-api-357b4e9c-27b4-4c57-875a-2032c7b53b53": Phase="Pending", Reason="", readiness=false. Elapsed: 12.218172597s
Jul 18 19:51:50.367: INFO: Pod "downward-api-357b4e9c-27b4-4c57-875a-2032c7b53b53": Phase="Succeeded", Reason="", readiness=false. Elapsed: 14.248988651s
STEP: Saw pod success
Jul 18 19:51:50.367: INFO: Pod "downward-api-357b4e9c-27b4-4c57-875a-2032c7b53b53" satisfied condition "Succeeded or Failed"
Jul 18 19:51:50.402: INFO: Trying to get logs from node 10.116.0.5 pod downward-api-357b4e9c-27b4-4c57-875a-2032c7b53b53 container dapi-container: <nil>
STEP: delete the pod
Jul 18 19:51:50.480: INFO: Waiting for pod downward-api-357b4e9c-27b4-4c57-875a-2032c7b53b53 to disappear
Jul 18 19:51:50.509: INFO: Pod downward-api-357b4e9c-27b4-4c57-875a-2032c7b53b53 no longer exists
[AfterEach] [sig-node] Downward API
  /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:186
... skipping 4 lines ...
• [SLOW TEST:14.678 seconds]
[sig-node] Downward API
/workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/common/node/framework.go:23
  should provide host IP and pod IP as an env var if pod uses host network [LinuxOnly]
  /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/common/node/downwardapi.go:109
------------------------------
{"msg":"PASSED [sig-node] Downward API should provide host IP and pod IP as an env var if pod uses host network [LinuxOnly]","total":-1,"completed":2,"skipped":3,"failed":0}

S
------------------------------
[BeforeEach] [Testpattern: Dynamic PV (default fs)] subPath
  /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/storage/framework/testsuite.go:51
Jul 18 19:51:50.620: INFO: Driver local doesn't support DynamicPV -- skipping
... skipping 53 lines ...
  /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:186
Jul 18 19:51:50.633: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready
STEP: Destroying namespace "configmap-3220" for this suite.

•
------------------------------
{"msg":"PASSED [sig-node] ConfigMap should update ConfigMap successfully","total":-1,"completed":2,"skipped":17,"failed":0}

SS
------------------------------
[BeforeEach] [sig-storage] CSI mock volume
  /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:185
STEP: Creating a kubernetes client
... skipping 101 lines ...
/workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/storage/utils/framework.go:23
  CSI attach test using mock driver
  /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/storage/csi_mock_volume.go:316
    should not require VolumeAttach for drivers without attachment
    /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/storage/csi_mock_volume.go:338
------------------------------
{"msg":"PASSED [sig-storage] CSI mock volume CSI attach test using mock driver should not require VolumeAttach for drivers without attachment","total":-1,"completed":2,"skipped":3,"failed":0}

SSSS
------------------------------
[BeforeEach] [sig-network] EndpointSlice
  /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:185
STEP: Creating a kubernetes client
... skipping 25 lines ...
  /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:186
Jul 18 19:51:51.548: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready
STEP: Destroying namespace "endpointslice-7385" for this suite.

•
------------------------------
{"msg":"PASSED [sig-network] EndpointSlice should support creating EndpointSlice API operations [Conformance]","total":-1,"completed":3,"skipped":19,"failed":0}

SSSSS
------------------------------
[BeforeEach] [Testpattern: Dynamic PV (block volmode)] volume-expand
  /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/storage/framework/testsuite.go:51
Jul 18 19:51:51.653: INFO: Driver emptydir doesn't support DynamicPV -- skipping
... skipping 92 lines ...
Jul 18 19:51:19.173: INFO: >>> kubeConfig: /root/.kube/config
STEP: Building a namespace api object, basename provisioning
STEP: Waiting for a default service account to be provisioned in namespace
[It] should support non-existent path
  /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/storage/testsuites/subpath.go:194
Jul 18 19:51:19.325: INFO: In-tree plugin kubernetes.io/host-path is not migrated, not validating any metrics
Jul 18 19:51:19.391: INFO: Waiting up to 5m0s for pod "hostpath-symlink-prep-provisioning-7175" in namespace "provisioning-7175" to be "Succeeded or Failed"
Jul 18 19:51:19.422: INFO: Pod "hostpath-symlink-prep-provisioning-7175": Phase="Pending", Reason="", readiness=false. Elapsed: 31.202426ms
Jul 18 19:51:21.455: INFO: Pod "hostpath-symlink-prep-provisioning-7175": Phase="Pending", Reason="", readiness=false. Elapsed: 2.063353512s
Jul 18 19:51:23.486: INFO: Pod "hostpath-symlink-prep-provisioning-7175": Phase="Pending", Reason="", readiness=false. Elapsed: 4.094814403s
Jul 18 19:51:25.524: INFO: Pod "hostpath-symlink-prep-provisioning-7175": Phase="Pending", Reason="", readiness=false. Elapsed: 6.132924416s
Jul 18 19:51:27.555: INFO: Pod "hostpath-symlink-prep-provisioning-7175": Phase="Succeeded", Reason="", readiness=false. Elapsed: 8.164162114s
STEP: Saw pod success
Jul 18 19:51:27.555: INFO: Pod "hostpath-symlink-prep-provisioning-7175" satisfied condition "Succeeded or Failed"
Jul 18 19:51:27.555: INFO: Deleting pod "hostpath-symlink-prep-provisioning-7175" in namespace "provisioning-7175"
Jul 18 19:51:27.599: INFO: Wait up to 5m0s for pod "hostpath-symlink-prep-provisioning-7175" to be fully deleted
Jul 18 19:51:27.629: INFO: Creating resource for inline volume
STEP: Creating pod pod-subpath-test-inlinevolume-lvxj
STEP: Creating a pod to test subpath
Jul 18 19:51:27.664: INFO: Waiting up to 5m0s for pod "pod-subpath-test-inlinevolume-lvxj" in namespace "provisioning-7175" to be "Succeeded or Failed"
Jul 18 19:51:27.695: INFO: Pod "pod-subpath-test-inlinevolume-lvxj": Phase="Pending", Reason="", readiness=false. Elapsed: 30.497535ms
Jul 18 19:51:29.725: INFO: Pod "pod-subpath-test-inlinevolume-lvxj": Phase="Pending", Reason="", readiness=false. Elapsed: 2.060828332s
Jul 18 19:51:31.758: INFO: Pod "pod-subpath-test-inlinevolume-lvxj": Phase="Pending", Reason="", readiness=false. Elapsed: 4.094160462s
Jul 18 19:51:33.789: INFO: Pod "pod-subpath-test-inlinevolume-lvxj": Phase="Pending", Reason="", readiness=false. Elapsed: 6.125044684s
Jul 18 19:51:35.822: INFO: Pod "pod-subpath-test-inlinevolume-lvxj": Phase="Pending", Reason="", readiness=false. Elapsed: 8.157257347s
Jul 18 19:51:37.854: INFO: Pod "pod-subpath-test-inlinevolume-lvxj": Phase="Pending", Reason="", readiness=false. Elapsed: 10.189668766s
Jul 18 19:51:39.885: INFO: Pod "pod-subpath-test-inlinevolume-lvxj": Phase="Succeeded", Reason="", readiness=false. Elapsed: 12.220925966s
STEP: Saw pod success
Jul 18 19:51:39.885: INFO: Pod "pod-subpath-test-inlinevolume-lvxj" satisfied condition "Succeeded or Failed"
Jul 18 19:51:39.915: INFO: Trying to get logs from node 10.116.0.7 pod pod-subpath-test-inlinevolume-lvxj container test-container-volume-inlinevolume-lvxj: <nil>
STEP: delete the pod
Jul 18 19:51:39.987: INFO: Waiting for pod pod-subpath-test-inlinevolume-lvxj to disappear
Jul 18 19:51:40.018: INFO: Pod pod-subpath-test-inlinevolume-lvxj no longer exists
STEP: Deleting pod pod-subpath-test-inlinevolume-lvxj
Jul 18 19:51:40.018: INFO: Deleting pod "pod-subpath-test-inlinevolume-lvxj" in namespace "provisioning-7175"
STEP: Deleting pod
Jul 18 19:51:40.049: INFO: Deleting pod "pod-subpath-test-inlinevolume-lvxj" in namespace "provisioning-7175"
Jul 18 19:51:40.116: INFO: Waiting up to 5m0s for pod "hostpath-symlink-prep-provisioning-7175" in namespace "provisioning-7175" to be "Succeeded or Failed"
Jul 18 19:51:40.146: INFO: Pod "hostpath-symlink-prep-provisioning-7175": Phase="Pending", Reason="", readiness=false. Elapsed: 30.204427ms
Jul 18 19:51:42.179: INFO: Pod "hostpath-symlink-prep-provisioning-7175": Phase="Pending", Reason="", readiness=false. Elapsed: 2.062755905s
Jul 18 19:51:44.211: INFO: Pod "hostpath-symlink-prep-provisioning-7175": Phase="Pending", Reason="", readiness=false. Elapsed: 4.094322754s
Jul 18 19:51:46.245: INFO: Pod "hostpath-symlink-prep-provisioning-7175": Phase="Pending", Reason="", readiness=false. Elapsed: 6.128781249s
Jul 18 19:51:48.277: INFO: Pod "hostpath-symlink-prep-provisioning-7175": Phase="Pending", Reason="", readiness=false. Elapsed: 8.160444026s
Jul 18 19:51:50.309: INFO: Pod "hostpath-symlink-prep-provisioning-7175": Phase="Pending", Reason="", readiness=false. Elapsed: 10.192238325s
Jul 18 19:51:52.340: INFO: Pod "hostpath-symlink-prep-provisioning-7175": Phase="Pending", Reason="", readiness=false. Elapsed: 12.223883559s
Jul 18 19:51:54.372: INFO: Pod "hostpath-symlink-prep-provisioning-7175": Phase="Pending", Reason="", readiness=false. Elapsed: 14.255236831s
Jul 18 19:51:56.405: INFO: Pod "hostpath-symlink-prep-provisioning-7175": Phase="Succeeded", Reason="", readiness=false. Elapsed: 16.288403638s
STEP: Saw pod success
Jul 18 19:51:56.405: INFO: Pod "hostpath-symlink-prep-provisioning-7175" satisfied condition "Succeeded or Failed"
Jul 18 19:51:56.405: INFO: Deleting pod "hostpath-symlink-prep-provisioning-7175" in namespace "provisioning-7175"
Jul 18 19:51:56.444: INFO: Wait up to 5m0s for pod "hostpath-symlink-prep-provisioning-7175" to be fully deleted
[AfterEach] [Testpattern: Inline-volume (default fs)] subPath
  /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:186
Jul 18 19:51:56.475: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready
STEP: Destroying namespace "provisioning-7175" for this suite.
... skipping 6 lines ...
  /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/storage/in_tree_volumes.go:58
    [Testpattern: Inline-volume (default fs)] subPath
    /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/storage/framework/testsuite.go:50
      should support non-existent path
      /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/storage/testsuites/subpath.go:194
------------------------------
{"msg":"PASSED [sig-storage] In-tree Volumes [Driver: hostPathSymlink] [Testpattern: Inline-volume (default fs)] subPath should support non-existent path","total":-1,"completed":1,"skipped":21,"failed":0}

SSSSSSS
------------------------------
[BeforeEach] [Testpattern: Dynamic PV (default fs)] capacity
  /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/storage/framework/testsuite.go:51
Jul 18 19:51:56.631: INFO: Driver emptydir doesn't support DynamicPV -- skipping
... skipping 57 lines ...
  /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:186
Jul 18 19:51:57.068: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready
STEP: Destroying namespace "events-5125" for this suite.

•
------------------------------
{"msg":"PASSED [sig-instrumentation] Events API should delete a collection of events [Conformance]","total":-1,"completed":2,"skipped":32,"failed":0}

SS
------------------------------
[BeforeEach] [Testpattern: Pre-provisioned PV (ext4)] volumes
  /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/storage/framework/testsuite.go:51
Jul 18 19:51:57.153: INFO: Driver hostPathSymlink doesn't support PreprovisionedPV -- skipping
... skipping 107 lines ...
STEP: Waiting for a default service account to be provisioned in namespace
[BeforeEach] [sig-storage] Downward API volume
  /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/common/storage/downwardapi_volume.go:41
[It] should set mode on item file [LinuxOnly] [NodeConformance] [Conformance]
  /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:630
STEP: Creating a pod to test downward API volume plugin
Jul 18 19:51:34.048: INFO: Waiting up to 5m0s for pod "downwardapi-volume-7ba485ed-ae6c-4fe5-8190-0e4bf7acd91b" in namespace "downward-api-5174" to be "Succeeded or Failed"
Jul 18 19:51:34.078: INFO: Pod "downwardapi-volume-7ba485ed-ae6c-4fe5-8190-0e4bf7acd91b": Phase="Pending", Reason="", readiness=false. Elapsed: 30.035836ms
Jul 18 19:51:36.111: INFO: Pod "downwardapi-volume-7ba485ed-ae6c-4fe5-8190-0e4bf7acd91b": Phase="Pending", Reason="", readiness=false. Elapsed: 2.062965173s
Jul 18 19:51:38.143: INFO: Pod "downwardapi-volume-7ba485ed-ae6c-4fe5-8190-0e4bf7acd91b": Phase="Pending", Reason="", readiness=false. Elapsed: 4.094856257s
Jul 18 19:51:40.174: INFO: Pod "downwardapi-volume-7ba485ed-ae6c-4fe5-8190-0e4bf7acd91b": Phase="Pending", Reason="", readiness=false. Elapsed: 6.125760912s
Jul 18 19:51:42.205: INFO: Pod "downwardapi-volume-7ba485ed-ae6c-4fe5-8190-0e4bf7acd91b": Phase="Pending", Reason="", readiness=false. Elapsed: 8.156844184s
Jul 18 19:51:44.236: INFO: Pod "downwardapi-volume-7ba485ed-ae6c-4fe5-8190-0e4bf7acd91b": Phase="Pending", Reason="", readiness=false. Elapsed: 10.187928337s
... skipping 2 lines ...
Jul 18 19:51:50.336: INFO: Pod "downwardapi-volume-7ba485ed-ae6c-4fe5-8190-0e4bf7acd91b": Phase="Pending", Reason="", readiness=false. Elapsed: 16.287996293s
Jul 18 19:51:52.371: INFO: Pod "downwardapi-volume-7ba485ed-ae6c-4fe5-8190-0e4bf7acd91b": Phase="Pending", Reason="", readiness=false. Elapsed: 18.322934791s
Jul 18 19:51:54.403: INFO: Pod "downwardapi-volume-7ba485ed-ae6c-4fe5-8190-0e4bf7acd91b": Phase="Pending", Reason="", readiness=false. Elapsed: 20.354647871s
Jul 18 19:51:56.435: INFO: Pod "downwardapi-volume-7ba485ed-ae6c-4fe5-8190-0e4bf7acd91b": Phase="Pending", Reason="", readiness=false. Elapsed: 22.38654058s
Jul 18 19:51:58.466: INFO: Pod "downwardapi-volume-7ba485ed-ae6c-4fe5-8190-0e4bf7acd91b": Phase="Succeeded", Reason="", readiness=false. Elapsed: 24.417653584s
STEP: Saw pod success
Jul 18 19:51:58.466: INFO: Pod "downwardapi-volume-7ba485ed-ae6c-4fe5-8190-0e4bf7acd91b" satisfied condition "Succeeded or Failed"
Jul 18 19:51:58.497: INFO: Trying to get logs from node 10.116.0.7 pod downwardapi-volume-7ba485ed-ae6c-4fe5-8190-0e4bf7acd91b container client-container: <nil>
STEP: delete the pod
Jul 18 19:51:58.578: INFO: Waiting for pod downwardapi-volume-7ba485ed-ae6c-4fe5-8190-0e4bf7acd91b to disappear
Jul 18 19:51:58.608: INFO: Pod downwardapi-volume-7ba485ed-ae6c-4fe5-8190-0e4bf7acd91b no longer exists
[AfterEach] [sig-storage] Downward API volume
  /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:186
... skipping 4 lines ...
• [SLOW TEST:24.845 seconds]
[sig-storage] Downward API volume
/workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/common/storage/framework.go:23
  should set mode on item file [LinuxOnly] [NodeConformance] [Conformance]
  /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:630
------------------------------
{"msg":"PASSED [sig-storage] Downward API volume should set mode on item file [LinuxOnly] [NodeConformance] [Conformance]","total":-1,"completed":3,"skipped":16,"failed":0}
[BeforeEach] [Testpattern: Dynamic PV (delayed binding)] topology
  /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/storage/framework/testsuite.go:51
Jul 18 19:51:58.715: INFO: Driver "csi-hostpath" does not support topology - skipping
[AfterEach] [Testpattern: Dynamic PV (delayed binding)] topology
  /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:186

... skipping 11 lines ...
      Driver "csi-hostpath" does not support topology - skipping

      /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/storage/testsuites/topology.go:92
------------------------------
SS
------------------------------
{"msg":"PASSED [sig-api-machinery] CustomResourcePublishOpenAPI [Privileged:ClusterAdmin] works for CRD without validation schema [Conformance]","total":-1,"completed":3,"skipped":22,"failed":0}
[BeforeEach] [sig-api-machinery] ServerSideApply
  /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:185
STEP: Creating a kubernetes client
Jul 18 19:51:58.423: INFO: >>> kubeConfig: /root/.kube/config
STEP: Building a namespace api object, basename apply
STEP: Waiting for a default service account to be provisioned in namespace
... skipping 7 lines ...
STEP: Destroying namespace "apply-6822" for this suite.
[AfterEach] [sig-api-machinery] ServerSideApply
  /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/apimachinery/apply.go:56

•
------------------------------
{"msg":"PASSED [sig-api-machinery] ServerSideApply should ignore conflict errors if force apply is used","total":-1,"completed":4,"skipped":22,"failed":0}

SS
------------------------------
[BeforeEach] [Testpattern: Pre-provisioned PV (default fs)] subPath
  /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/storage/framework/testsuite.go:51
[BeforeEach] [Testpattern: Pre-provisioned PV (default fs)] subPath
... skipping 17 lines ...
Jul 18 19:51:34.802: INFO: PersistentVolumeClaim pvc-ld2lf found but phase is Pending instead of Bound.
Jul 18 19:51:36.836: INFO: PersistentVolumeClaim pvc-ld2lf found and phase=Bound (4.099768243s)
Jul 18 19:51:36.836: INFO: Waiting up to 3m0s for PersistentVolume local-bc4rt to have phase Bound
Jul 18 19:51:36.867: INFO: PersistentVolume local-bc4rt found and phase=Bound (31.322246ms)
STEP: Creating pod pod-subpath-test-preprovisionedpv-d5qq
STEP: Creating a pod to test subpath
Jul 18 19:51:36.974: INFO: Waiting up to 5m0s for pod "pod-subpath-test-preprovisionedpv-d5qq" in namespace "provisioning-4722" to be "Succeeded or Failed"
Jul 18 19:51:37.005: INFO: Pod "pod-subpath-test-preprovisionedpv-d5qq": Phase="Pending", Reason="", readiness=false. Elapsed: 31.015972ms
Jul 18 19:51:39.040: INFO: Pod "pod-subpath-test-preprovisionedpv-d5qq": Phase="Pending", Reason="", readiness=false. Elapsed: 2.065933286s
Jul 18 19:51:41.076: INFO: Pod "pod-subpath-test-preprovisionedpv-d5qq": Phase="Pending", Reason="", readiness=false. Elapsed: 4.101895075s
Jul 18 19:51:43.109: INFO: Pod "pod-subpath-test-preprovisionedpv-d5qq": Phase="Pending", Reason="", readiness=false. Elapsed: 6.135396178s
Jul 18 19:51:45.142: INFO: Pod "pod-subpath-test-preprovisionedpv-d5qq": Phase="Pending", Reason="", readiness=false. Elapsed: 8.167731408s
Jul 18 19:51:47.173: INFO: Pod "pod-subpath-test-preprovisionedpv-d5qq": Phase="Pending", Reason="", readiness=false. Elapsed: 10.199276655s
Jul 18 19:51:49.205: INFO: Pod "pod-subpath-test-preprovisionedpv-d5qq": Phase="Pending", Reason="", readiness=false. Elapsed: 12.230953423s
Jul 18 19:51:51.237: INFO: Pod "pod-subpath-test-preprovisionedpv-d5qq": Phase="Pending", Reason="", readiness=false. Elapsed: 14.263122485s
Jul 18 19:51:53.269: INFO: Pod "pod-subpath-test-preprovisionedpv-d5qq": Phase="Pending", Reason="", readiness=false. Elapsed: 16.295300912s
Jul 18 19:51:55.302: INFO: Pod "pod-subpath-test-preprovisionedpv-d5qq": Phase="Pending", Reason="", readiness=false. Elapsed: 18.327842985s
Jul 18 19:51:57.333: INFO: Pod "pod-subpath-test-preprovisionedpv-d5qq": Phase="Pending", Reason="", readiness=false. Elapsed: 20.358847362s
Jul 18 19:51:59.366: INFO: Pod "pod-subpath-test-preprovisionedpv-d5qq": Phase="Succeeded", Reason="", readiness=false. Elapsed: 22.391877146s
STEP: Saw pod success
Jul 18 19:51:59.366: INFO: Pod "pod-subpath-test-preprovisionedpv-d5qq" satisfied condition "Succeeded or Failed"
Jul 18 19:51:59.402: INFO: Trying to get logs from node 10.116.0.5 pod pod-subpath-test-preprovisionedpv-d5qq container test-container-subpath-preprovisionedpv-d5qq: <nil>
STEP: delete the pod
Jul 18 19:51:59.481: INFO: Waiting for pod pod-subpath-test-preprovisionedpv-d5qq to disappear
Jul 18 19:51:59.512: INFO: Pod pod-subpath-test-preprovisionedpv-d5qq no longer exists
STEP: Deleting pod pod-subpath-test-preprovisionedpv-d5qq
Jul 18 19:51:59.512: INFO: Deleting pod "pod-subpath-test-preprovisionedpv-d5qq" in namespace "provisioning-4722"
... skipping 22 lines ...
  /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/storage/in_tree_volumes.go:58
    [Testpattern: Pre-provisioned PV (default fs)] subPath
    /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/storage/framework/testsuite.go:50
      should support existing single file [LinuxOnly]
      /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/storage/testsuites/subpath.go:219
------------------------------
{"msg":"PASSED [sig-storage] In-tree Volumes [Driver: local][LocalVolumeType: tmpfs] [Testpattern: Pre-provisioned PV (default fs)] subPath should support existing single file [LinuxOnly]","total":-1,"completed":4,"skipped":42,"failed":0}

SSSS
------------------------------
[BeforeEach] [sig-api-machinery] Server request timeout
  /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:185
STEP: Creating a kubernetes client
... skipping 6 lines ...
  /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:186
Jul 18 19:52:00.599: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready
STEP: Destroying namespace "request-timeout-4049" for this suite.

•
------------------------------
{"msg":"PASSED [sig-api-machinery] Server request timeout the request should be served with a default timeout if the specified timeout in the request URL exceeds maximum allowed","total":-1,"completed":5,"skipped":46,"failed":0}

S
------------------------------
[BeforeEach] [Testpattern: Dynamic PV (block volmode)(allowExpansion)] volume-expand
  /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/storage/framework/testsuite.go:51
Jul 18 19:52:00.680: INFO: Driver local doesn't support DynamicPV -- skipping
... skipping 30 lines ...
  /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:186
Jul 18 19:52:01.239: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready
STEP: Destroying namespace "health-2628" for this suite.

•
------------------------------
{"msg":"PASSED [sig-api-machinery] health handlers should contain necessary checks","total":-1,"completed":6,"skipped":48,"failed":0}

SS
------------------------------
[BeforeEach] [Testpattern: Pre-provisioned PV (default fs)] subPath
  /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/storage/framework/testsuite.go:51
[BeforeEach] [Testpattern: Pre-provisioned PV (default fs)] subPath
... skipping 19 lines ...
Jul 18 19:51:21.934: INFO: PersistentVolumeClaim pvc-px9nz found but phase is Pending instead of Bound.
Jul 18 19:51:23.965: INFO: PersistentVolumeClaim pvc-px9nz found and phase=Bound (10.19091551s)
Jul 18 19:51:23.966: INFO: Waiting up to 3m0s for PersistentVolume local-vdphf to have phase Bound
Jul 18 19:51:23.997: INFO: PersistentVolume local-vdphf found and phase=Bound (30.952067ms)
STEP: Creating pod pod-subpath-test-preprovisionedpv-nklq
STEP: Creating a pod to test subpath
Jul 18 19:51:24.097: INFO: Waiting up to 5m0s for pod "pod-subpath-test-preprovisionedpv-nklq" in namespace "provisioning-543" to be "Succeeded or Failed"
Jul 18 19:51:24.127: INFO: Pod "pod-subpath-test-preprovisionedpv-nklq": Phase="Pending", Reason="", readiness=false. Elapsed: 30.125546ms
Jul 18 19:51:26.160: INFO: Pod "pod-subpath-test-preprovisionedpv-nklq": Phase="Pending", Reason="", readiness=false. Elapsed: 2.062617568s
Jul 18 19:51:28.192: INFO: Pod "pod-subpath-test-preprovisionedpv-nklq": Phase="Pending", Reason="", readiness=false. Elapsed: 4.095225186s
Jul 18 19:51:30.228: INFO: Pod "pod-subpath-test-preprovisionedpv-nklq": Phase="Pending", Reason="", readiness=false. Elapsed: 6.130399481s
Jul 18 19:51:32.261: INFO: Pod "pod-subpath-test-preprovisionedpv-nklq": Phase="Pending", Reason="", readiness=false. Elapsed: 8.163400811s
Jul 18 19:51:34.295: INFO: Pod "pod-subpath-test-preprovisionedpv-nklq": Phase="Pending", Reason="", readiness=false. Elapsed: 10.197454196s
... skipping 8 lines ...
Jul 18 19:51:52.652: INFO: Pod "pod-subpath-test-preprovisionedpv-nklq": Phase="Pending", Reason="", readiness=false. Elapsed: 28.554843329s
Jul 18 19:51:54.686: INFO: Pod "pod-subpath-test-preprovisionedpv-nklq": Phase="Pending", Reason="", readiness=false. Elapsed: 30.588581586s
Jul 18 19:51:56.723: INFO: Pod "pod-subpath-test-preprovisionedpv-nklq": Phase="Pending", Reason="", readiness=false. Elapsed: 32.626006568s
Jul 18 19:51:58.759: INFO: Pod "pod-subpath-test-preprovisionedpv-nklq": Phase="Pending", Reason="", readiness=false. Elapsed: 34.661754849s
Jul 18 19:52:00.792: INFO: Pod "pod-subpath-test-preprovisionedpv-nklq": Phase="Succeeded", Reason="", readiness=false. Elapsed: 36.694741873s
STEP: Saw pod success
Jul 18 19:52:00.792: INFO: Pod "pod-subpath-test-preprovisionedpv-nklq" satisfied condition "Succeeded or Failed"
Jul 18 19:52:00.823: INFO: Trying to get logs from node 10.116.0.5 pod pod-subpath-test-preprovisionedpv-nklq container test-container-volume-preprovisionedpv-nklq: <nil>
STEP: delete the pod
Jul 18 19:52:00.903: INFO: Waiting for pod pod-subpath-test-preprovisionedpv-nklq to disappear
Jul 18 19:52:00.934: INFO: Pod pod-subpath-test-preprovisionedpv-nklq no longer exists
STEP: Deleting pod pod-subpath-test-preprovisionedpv-nklq
Jul 18 19:52:00.934: INFO: Deleting pod "pod-subpath-test-preprovisionedpv-nklq" in namespace "provisioning-543"
... skipping 19 lines ...
  /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/storage/in_tree_volumes.go:58
    [Testpattern: Pre-provisioned PV (default fs)] subPath
    /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/storage/framework/testsuite.go:50
      should support non-existent path
      /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/storage/testsuites/subpath.go:194
------------------------------
{"msg":"PASSED [sig-storage] In-tree Volumes [Driver: local][LocalVolumeType: dir-bindmounted] [Testpattern: Pre-provisioned PV (default fs)] subPath should support non-existent path","total":-1,"completed":3,"skipped":10,"failed":0}
[BeforeEach] [Testpattern: Pre-provisioned PV (block volmode)] volumeMode
  /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/storage/framework/testsuite.go:51
Jul 18 19:52:01.483: INFO: Driver csi-hostpath doesn't support PreprovisionedPV -- skipping
[AfterEach] [Testpattern: Pre-provisioned PV (block volmode)] volumeMode
  /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:186

... skipping 76 lines ...
Jul 18 19:51:57.064: INFO: Running '/tmp/kubectl2220469345/kubectl --server=https://143.244.215.182 --kubeconfig=/root/.kube/config --namespace=crd-publish-openapi-4099 explain e2e-test-crd-publish-openapi-9733-crds.spec'
Jul 18 19:51:57.479: INFO: stderr: ""
Jul 18 19:51:57.479: INFO: stdout: "KIND:     E2e-test-crd-publish-openapi-9733-crd\nVERSION:  crd-publish-openapi-test-foo.example.com/v1\n\nRESOURCE: spec <Object>\n\nDESCRIPTION:\n     Specification of Foo\n\nFIELDS:\n   bars\t<[]Object>\n     List of Bars and their specs.\n\n"
Jul 18 19:51:57.480: INFO: Running '/tmp/kubectl2220469345/kubectl --server=https://143.244.215.182 --kubeconfig=/root/.kube/config --namespace=crd-publish-openapi-4099 explain e2e-test-crd-publish-openapi-9733-crds.spec.bars'
Jul 18 19:51:57.925: INFO: stderr: ""
Jul 18 19:51:57.925: INFO: stdout: "KIND:     E2e-test-crd-publish-openapi-9733-crd\nVERSION:  crd-publish-openapi-test-foo.example.com/v1\n\nRESOURCE: bars <[]Object>\n\nDESCRIPTION:\n     List of Bars and their specs.\n\nFIELDS:\n   age\t<string>\n     Age of Bar.\n\n   bazs\t<[]string>\n     List of Bazs.\n\n   name\t<string> -required-\n     Name of Bar.\n\n"
STEP: kubectl explain works to return error when explain is called on property that doesn't exist
Jul 18 19:51:57.925: INFO: Running '/tmp/kubectl2220469345/kubectl --server=https://143.244.215.182 --kubeconfig=/root/.kube/config --namespace=crd-publish-openapi-4099 explain e2e-test-crd-publish-openapi-9733-crds.spec.bars2'
Jul 18 19:51:58.356: INFO: rc: 1
[AfterEach] [sig-api-machinery] CustomResourcePublishOpenAPI [Privileged:ClusterAdmin]
  /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:186
Jul 18 19:52:01.908: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready
STEP: Destroying namespace "crd-publish-openapi-4099" for this suite.
... skipping 2 lines ...
• [SLOW TEST:17.730 seconds]
[sig-api-machinery] CustomResourcePublishOpenAPI [Privileged:ClusterAdmin]
/workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/apimachinery/framework.go:23
  works for CRD with validation schema [Conformance]
  /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:630
------------------------------
{"msg":"PASSED [sig-api-machinery] CustomResourcePublishOpenAPI [Privileged:ClusterAdmin] works for CRD with validation schema [Conformance]","total":-1,"completed":3,"skipped":7,"failed":0}

S
------------------------------
[BeforeEach] [Testpattern: Inline-volume (default fs)] subPath
  /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/storage/framework/testsuite.go:51
Jul 18 19:52:02.014: INFO: Driver local doesn't support InlineVolume -- skipping
... skipping 40 lines ...
• [SLOW TEST:61.371 seconds]
[sig-node] Probing container
/workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/common/node/framework.go:23
  should be restarted with an exec liveness probe with timeout [MinimumKubeletVersion:1.20] [NodeConformance]
  /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/common/node/container_probe.go:217
------------------------------
{"msg":"PASSED [sig-node] Probing container should be restarted with an exec liveness probe with timeout [MinimumKubeletVersion:1.20] [NodeConformance]","total":-1,"completed":3,"skipped":8,"failed":0}

SS
------------------------------
[BeforeEach] [Testpattern: Inline-volume (ext4)] volumes
  /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/storage/framework/testsuite.go:51
Jul 18 19:52:02.168: INFO: Driver local doesn't support InlineVolume -- skipping
... skipping 68 lines ...
Jul 18 19:51:21.041: INFO: PersistentVolumeClaim pvc-s2z9l found but phase is Pending instead of Bound.
Jul 18 19:51:23.073: INFO: PersistentVolumeClaim pvc-s2z9l found and phase=Bound (14.25172384s)
Jul 18 19:51:23.073: INFO: Waiting up to 3m0s for PersistentVolume local-q8vl6 to have phase Bound
Jul 18 19:51:23.103: INFO: PersistentVolume local-q8vl6 found and phase=Bound (30.57369ms)
STEP: Creating pod pod-subpath-test-preprovisionedpv-7zsv
STEP: Creating a pod to test atomic-volume-subpath
Jul 18 19:51:23.205: INFO: Waiting up to 5m0s for pod "pod-subpath-test-preprovisionedpv-7zsv" in namespace "provisioning-1347" to be "Succeeded or Failed"
Jul 18 19:51:23.236: INFO: Pod "pod-subpath-test-preprovisionedpv-7zsv": Phase="Pending", Reason="", readiness=false. Elapsed: 31.326087ms
Jul 18 19:51:25.271: INFO: Pod "pod-subpath-test-preprovisionedpv-7zsv": Phase="Pending", Reason="", readiness=false. Elapsed: 2.065945872s
Jul 18 19:51:27.301: INFO: Pod "pod-subpath-test-preprovisionedpv-7zsv": Phase="Pending", Reason="", readiness=false. Elapsed: 4.096486288s
Jul 18 19:51:29.333: INFO: Pod "pod-subpath-test-preprovisionedpv-7zsv": Phase="Pending", Reason="", readiness=false. Elapsed: 6.128042279s
Jul 18 19:51:31.364: INFO: Pod "pod-subpath-test-preprovisionedpv-7zsv": Phase="Pending", Reason="", readiness=false. Elapsed: 8.159334004s
Jul 18 19:51:33.394: INFO: Pod "pod-subpath-test-preprovisionedpv-7zsv": Phase="Pending", Reason="", readiness=false. Elapsed: 10.189458766s
... skipping 9 lines ...
Jul 18 19:51:53.718: INFO: Pod "pod-subpath-test-preprovisionedpv-7zsv": Phase="Running", Reason="", readiness=true. Elapsed: 30.513768136s
Jul 18 19:51:55.753: INFO: Pod "pod-subpath-test-preprovisionedpv-7zsv": Phase="Running", Reason="", readiness=true. Elapsed: 32.548463909s
Jul 18 19:51:57.785: INFO: Pod "pod-subpath-test-preprovisionedpv-7zsv": Phase="Running", Reason="", readiness=true. Elapsed: 34.580359095s
Jul 18 19:51:59.838: INFO: Pod "pod-subpath-test-preprovisionedpv-7zsv": Phase="Running", Reason="", readiness=true. Elapsed: 36.633329364s
Jul 18 19:52:01.870: INFO: Pod "pod-subpath-test-preprovisionedpv-7zsv": Phase="Succeeded", Reason="", readiness=false. Elapsed: 38.665408709s
STEP: Saw pod success
Jul 18 19:52:01.870: INFO: Pod "pod-subpath-test-preprovisionedpv-7zsv" satisfied condition "Succeeded or Failed"
Jul 18 19:52:01.900: INFO: Trying to get logs from node 10.116.0.5 pod pod-subpath-test-preprovisionedpv-7zsv container test-container-subpath-preprovisionedpv-7zsv: <nil>
STEP: delete the pod
Jul 18 19:52:01.975: INFO: Waiting for pod pod-subpath-test-preprovisionedpv-7zsv to disappear
Jul 18 19:52:02.005: INFO: Pod pod-subpath-test-preprovisionedpv-7zsv no longer exists
STEP: Deleting pod pod-subpath-test-preprovisionedpv-7zsv
Jul 18 19:52:02.005: INFO: Deleting pod "pod-subpath-test-preprovisionedpv-7zsv" in namespace "provisioning-1347"
... skipping 22 lines ...
  /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/storage/in_tree_volumes.go:58
    [Testpattern: Pre-provisioned PV (default fs)] subPath
    /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/storage/framework/testsuite.go:50
      should support file as subpath [LinuxOnly]
      /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/storage/testsuites/subpath.go:230
------------------------------
{"msg":"PASSED [sig-storage] In-tree Volumes [Driver: local][LocalVolumeType: tmpfs] [Testpattern: Pre-provisioned PV (default fs)] subPath should support file as subpath [LinuxOnly]","total":-1,"completed":3,"skipped":38,"failed":0}

SS
------------------------------
[BeforeEach] [Testpattern: Pre-provisioned PV (default fs)] subPath
  /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/storage/framework/testsuite.go:51
[BeforeEach] [Testpattern: Pre-provisioned PV (default fs)] subPath
... skipping 16 lines ...
Jul 18 19:51:51.465: INFO: PersistentVolumeClaim pvc-pkr7t found but phase is Pending instead of Bound.
Jul 18 19:51:53.498: INFO: PersistentVolumeClaim pvc-pkr7t found and phase=Bound (2.063048658s)
Jul 18 19:51:53.498: INFO: Waiting up to 3m0s for PersistentVolume local-9m9sw to have phase Bound
Jul 18 19:51:53.529: INFO: PersistentVolume local-9m9sw found and phase=Bound (30.512179ms)
STEP: Creating pod pod-subpath-test-preprovisionedpv-vn4n
STEP: Creating a pod to test subpath
Jul 18 19:51:53.636: INFO: Waiting up to 5m0s for pod "pod-subpath-test-preprovisionedpv-vn4n" in namespace "provisioning-6875" to be "Succeeded or Failed"
Jul 18 19:51:53.666: INFO: Pod "pod-subpath-test-preprovisionedpv-vn4n": Phase="Pending", Reason="", readiness=false. Elapsed: 30.119709ms
Jul 18 19:51:55.702: INFO: Pod "pod-subpath-test-preprovisionedpv-vn4n": Phase="Pending", Reason="", readiness=false. Elapsed: 2.066336937s
Jul 18 19:51:57.734: INFO: Pod "pod-subpath-test-preprovisionedpv-vn4n": Phase="Pending", Reason="", readiness=false. Elapsed: 4.098784478s
Jul 18 19:51:59.766: INFO: Pod "pod-subpath-test-preprovisionedpv-vn4n": Phase="Pending", Reason="", readiness=false. Elapsed: 6.130585768s
Jul 18 19:52:01.797: INFO: Pod "pod-subpath-test-preprovisionedpv-vn4n": Phase="Pending", Reason="", readiness=false. Elapsed: 8.161020431s
Jul 18 19:52:03.828: INFO: Pod "pod-subpath-test-preprovisionedpv-vn4n": Phase="Succeeded", Reason="", readiness=false. Elapsed: 10.192075644s
STEP: Saw pod success
Jul 18 19:52:03.828: INFO: Pod "pod-subpath-test-preprovisionedpv-vn4n" satisfied condition "Succeeded or Failed"
Jul 18 19:52:03.858: INFO: Trying to get logs from node 10.116.0.5 pod pod-subpath-test-preprovisionedpv-vn4n container test-container-subpath-preprovisionedpv-vn4n: <nil>
STEP: delete the pod
Jul 18 19:52:03.939: INFO: Waiting for pod pod-subpath-test-preprovisionedpv-vn4n to disappear
Jul 18 19:52:03.969: INFO: Pod pod-subpath-test-preprovisionedpv-vn4n no longer exists
STEP: Deleting pod pod-subpath-test-preprovisionedpv-vn4n
Jul 18 19:52:03.969: INFO: Deleting pod "pod-subpath-test-preprovisionedpv-vn4n" in namespace "provisioning-6875"
... skipping 22 lines ...
  /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/storage/in_tree_volumes.go:58
    [Testpattern: Pre-provisioned PV (default fs)] subPath
    /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/storage/framework/testsuite.go:50
      should support readOnly file specified in the volumeMount [LinuxOnly]
      /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/storage/testsuites/subpath.go:379
------------------------------
{"msg":"PASSED [sig-storage] In-tree Volumes [Driver: local][LocalVolumeType: tmpfs] [Testpattern: Pre-provisioned PV (default fs)] subPath should support readOnly file specified in the volumeMount [LinuxOnly]","total":-1,"completed":3,"skipped":13,"failed":0}

SSS
------------------------------
[BeforeEach] [Testpattern: Dynamic PV (default fs)] fsgroupchangepolicy
  /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/storage/framework/testsuite.go:51
Jul 18 19:52:04.764: INFO: Driver local doesn't support DynamicPV -- skipping
... skipping 39 lines ...
Jul 18 19:51:50.208: INFO: PersistentVolumeClaim pvc-w2crx found but phase is Pending instead of Bound.
Jul 18 19:51:52.241: INFO: PersistentVolumeClaim pvc-w2crx found and phase=Bound (8.161305678s)
Jul 18 19:51:52.241: INFO: Waiting up to 3m0s for PersistentVolume local-r5kxq to have phase Bound
Jul 18 19:51:52.273: INFO: PersistentVolume local-r5kxq found and phase=Bound (31.931858ms)
STEP: Creating pod exec-volume-test-preprovisionedpv-qz6g
STEP: Creating a pod to test exec-volume-test
Jul 18 19:51:52.371: INFO: Waiting up to 5m0s for pod "exec-volume-test-preprovisionedpv-qz6g" in namespace "volume-4743" to be "Succeeded or Failed"
Jul 18 19:51:52.403: INFO: Pod "exec-volume-test-preprovisionedpv-qz6g": Phase="Pending", Reason="", readiness=false. Elapsed: 31.739143ms
Jul 18 19:51:54.440: INFO: Pod "exec-volume-test-preprovisionedpv-qz6g": Phase="Pending", Reason="", readiness=false. Elapsed: 2.068490039s
Jul 18 19:51:56.474: INFO: Pod "exec-volume-test-preprovisionedpv-qz6g": Phase="Pending", Reason="", readiness=false. Elapsed: 4.102411927s
Jul 18 19:51:58.505: INFO: Pod "exec-volume-test-preprovisionedpv-qz6g": Phase="Pending", Reason="", readiness=false. Elapsed: 6.134134442s
Jul 18 19:52:00.539: INFO: Pod "exec-volume-test-preprovisionedpv-qz6g": Phase="Pending", Reason="", readiness=false. Elapsed: 8.167276853s
Jul 18 19:52:02.572: INFO: Pod "exec-volume-test-preprovisionedpv-qz6g": Phase="Pending", Reason="", readiness=false. Elapsed: 10.200200888s
Jul 18 19:52:04.606: INFO: Pod "exec-volume-test-preprovisionedpv-qz6g": Phase="Pending", Reason="", readiness=false. Elapsed: 12.235126088s
Jul 18 19:52:06.639: INFO: Pod "exec-volume-test-preprovisionedpv-qz6g": Phase="Succeeded", Reason="", readiness=false. Elapsed: 14.267500317s
STEP: Saw pod success
Jul 18 19:52:06.639: INFO: Pod "exec-volume-test-preprovisionedpv-qz6g" satisfied condition "Succeeded or Failed"
Jul 18 19:52:06.670: INFO: Trying to get logs from node 10.116.0.5 pod exec-volume-test-preprovisionedpv-qz6g container exec-container-preprovisionedpv-qz6g: <nil>
STEP: delete the pod
Jul 18 19:52:06.754: INFO: Waiting for pod exec-volume-test-preprovisionedpv-qz6g to disappear
Jul 18 19:52:06.785: INFO: Pod exec-volume-test-preprovisionedpv-qz6g no longer exists
STEP: Deleting pod exec-volume-test-preprovisionedpv-qz6g
Jul 18 19:52:06.785: INFO: Deleting pod "exec-volume-test-preprovisionedpv-qz6g" in namespace "volume-4743"
... skipping 17 lines ...
  /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/storage/in_tree_volumes.go:58
    [Testpattern: Pre-provisioned PV (default fs)] volumes
    /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/storage/framework/testsuite.go:50
      should allow exec of files on the volume
      /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/storage/testsuites/volumes.go:196
------------------------------
{"msg":"PASSED [sig-storage] In-tree Volumes [Driver: local][LocalVolumeType: dir-bindmounted] [Testpattern: Pre-provisioned PV (default fs)] volumes should allow exec of files on the volume","total":-1,"completed":4,"skipped":33,"failed":0}

S
------------------------------
{"msg":"PASSED [sig-cli] Kubectl Port forwarding With a server listening on 0.0.0.0 that expects NO client request should support a client that connects, sends DATA, and disconnects","total":-1,"completed":1,"skipped":16,"failed":0}
[BeforeEach] [sig-storage] PersistentVolumes-local 
  /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:185
STEP: Creating a kubernetes client
Jul 18 19:51:08.814: INFO: >>> kubeConfig: /root/.kube/config
STEP: Building a namespace api object, basename persistent-local-volumes-test
STEP: Waiting for a default service account to be provisioned in namespace
... skipping 74 lines ...
  /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/storage/persistent_volumes-local.go:192
    Two pods mounting a local volume at the same time
    /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/storage/persistent_volumes-local.go:248
      should be able to write from pod1 and read from pod2
      /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/storage/persistent_volumes-local.go:249
------------------------------
{"msg":"PASSED [sig-storage] PersistentVolumes-local  [Volume type: blockfswithformat] Two pods mounting a local volume at the same time should be able to write from pod1 and read from pod2","total":-1,"completed":2,"skipped":16,"failed":0}

SSSS
------------------------------
[BeforeEach] [Testpattern: Inline-volume (default fs)] volumes
  /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/storage/framework/testsuite.go:51
Jul 18 19:52:07.340: INFO: Only supported for node OS distro [gci ubuntu custom] (not debian)
... skipping 18 lines ...
[BeforeEach] [sig-node] Secrets
  /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:185
STEP: Creating a kubernetes client
Jul 18 19:52:07.316: INFO: >>> kubeConfig: /root/.kube/config
STEP: Building a namespace api object, basename secrets
STEP: Waiting for a default service account to be provisioned in namespace
[It] should fail to create secret due to empty secret key [Conformance]
  /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:630
STEP: Creating projection with secret that has name secret-emptykey-test-0ed79c90-3e08-4a3b-a724-c7860b129555
[AfterEach] [sig-node] Secrets
  /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:186
Jul 18 19:52:07.508: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready
STEP: Destroying namespace "secrets-4190" for this suite.

•
------------------------------
{"msg":"PASSED [sig-node] Secrets should fail to create secret due to empty secret key [Conformance]","total":-1,"completed":5,"skipped":34,"failed":0}

S
------------------------------
[BeforeEach] [Testpattern: Dynamic PV (default fs)] subPath
  /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/storage/framework/testsuite.go:51
Jul 18 19:52:07.711: INFO: Only supported for providers [gce gke] (not skeleton)
... skipping 126 lines ...
Jul 18 19:51:50.117: INFO: PersistentVolumeClaim pvc-r99pb found but phase is Pending instead of Bound.
Jul 18 19:51:52.154: INFO: PersistentVolumeClaim pvc-r99pb found and phase=Bound (6.129339249s)
Jul 18 19:51:52.154: INFO: Waiting up to 3m0s for PersistentVolume local-nn67v to have phase Bound
Jul 18 19:51:52.187: INFO: PersistentVolume local-nn67v found and phase=Bound (32.00917ms)
STEP: Creating pod pod-subpath-test-preprovisionedpv-4t77
STEP: Creating a pod to test subpath
Jul 18 19:51:52.286: INFO: Waiting up to 5m0s for pod "pod-subpath-test-preprovisionedpv-4t77" in namespace "provisioning-5994" to be "Succeeded or Failed"
Jul 18 19:51:52.317: INFO: Pod "pod-subpath-test-preprovisionedpv-4t77": Phase="Pending", Reason="", readiness=false. Elapsed: 31.346785ms
Jul 18 19:51:54.349: INFO: Pod "pod-subpath-test-preprovisionedpv-4t77": Phase="Pending", Reason="", readiness=false. Elapsed: 2.063133918s
Jul 18 19:51:56.383: INFO: Pod "pod-subpath-test-preprovisionedpv-4t77": Phase="Pending", Reason="", readiness=false. Elapsed: 4.097529844s
Jul 18 19:51:58.416: INFO: Pod "pod-subpath-test-preprovisionedpv-4t77": Phase="Pending", Reason="", readiness=false. Elapsed: 6.129749408s
Jul 18 19:52:00.448: INFO: Pod "pod-subpath-test-preprovisionedpv-4t77": Phase="Pending", Reason="", readiness=false. Elapsed: 8.162567002s
Jul 18 19:52:02.480: INFO: Pod "pod-subpath-test-preprovisionedpv-4t77": Phase="Pending", Reason="", readiness=false. Elapsed: 10.194250871s
Jul 18 19:52:04.512: INFO: Pod "pod-subpath-test-preprovisionedpv-4t77": Phase="Pending", Reason="", readiness=false. Elapsed: 12.226195626s
Jul 18 19:52:06.565: INFO: Pod "pod-subpath-test-preprovisionedpv-4t77": Phase="Pending", Reason="", readiness=false. Elapsed: 14.279323671s
Jul 18 19:52:08.599: INFO: Pod "pod-subpath-test-preprovisionedpv-4t77": Phase="Succeeded", Reason="", readiness=false. Elapsed: 16.31319178s
STEP: Saw pod success
Jul 18 19:52:08.599: INFO: Pod "pod-subpath-test-preprovisionedpv-4t77" satisfied condition "Succeeded or Failed"
Jul 18 19:52:08.633: INFO: Trying to get logs from node 10.116.0.5 pod pod-subpath-test-preprovisionedpv-4t77 container test-container-subpath-preprovisionedpv-4t77: <nil>
STEP: delete the pod
Jul 18 19:52:08.710: INFO: Waiting for pod pod-subpath-test-preprovisionedpv-4t77 to disappear
Jul 18 19:52:08.741: INFO: Pod pod-subpath-test-preprovisionedpv-4t77 no longer exists
STEP: Deleting pod pod-subpath-test-preprovisionedpv-4t77
Jul 18 19:52:08.741: INFO: Deleting pod "pod-subpath-test-preprovisionedpv-4t77" in namespace "provisioning-5994"
... skipping 19 lines ...
  /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/storage/in_tree_volumes.go:58
    [Testpattern: Pre-provisioned PV (default fs)] subPath
    /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/storage/framework/testsuite.go:50
      should support existing single file [LinuxOnly]
      /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/storage/testsuites/subpath.go:219
------------------------------
{"msg":"PASSED [sig-storage] In-tree Volumes [Driver: local][LocalVolumeType: dir-bindmounted] [Testpattern: Pre-provisioned PV (default fs)] subPath should support existing single file [LinuxOnly]","total":-1,"completed":6,"skipped":41,"failed":0}

S
------------------------------
[BeforeEach] [Testpattern: Inline-volume (default fs)] subPath
  /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/storage/framework/testsuite.go:51
Jul 18 19:52:09.296: INFO: Driver local doesn't support InlineVolume -- skipping
... skipping 130 lines ...
STEP: Building a namespace api object, basename pvc-protection
STEP: Waiting for a default service account to be provisioned in namespace
[BeforeEach] [sig-storage] PVC Protection
  /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/storage/pvc_protection.go:72
Jul 18 19:52:12.069: INFO: Waiting up to 30m0s for all (but 0) nodes to be schedulable
STEP: Creating a PVC
Jul 18 19:52:12.130: INFO: error finding default storageClass : No default storage class found
[AfterEach] [sig-storage] PVC Protection
  /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:186
Jul 18 19:52:12.132: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready
STEP: Destroying namespace "pvc-protection-827" for this suite.
[AfterEach] [sig-storage] PVC Protection
  /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/storage/pvc_protection.go:108
... skipping 2 lines ...
S [SKIPPING] in Spec Setup (BeforeEach) [0.286 seconds]
[sig-storage] PVC Protection
/workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/storage/utils/framework.go:23
  Verify that scheduling of a pod that uses PVC that is being deleted fails and the pod becomes Unschedulable [BeforeEach]
  /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/storage/pvc_protection.go:145

  error finding default storageClass : No default storage class found

  /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/pv/pv.go:819
------------------------------
SS
------------------------------
[BeforeEach] [sig-api-machinery] ServerSideApply
... skipping 12 lines ...
STEP: Destroying namespace "apply-1579" for this suite.
[AfterEach] [sig-api-machinery] ServerSideApply
  /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/apimachinery/apply.go:56

•
------------------------------
{"msg":"PASSED [sig-api-machinery] ServerSideApply should create an applied object if it does not already exist","total":-1,"completed":3,"skipped":19,"failed":0}

SSS
------------------------------
[BeforeEach] [Testpattern: Dynamic PV (block volmode)(allowExpansion)] volume-expand
  /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/storage/framework/testsuite.go:51
Jul 18 19:52:12.814: INFO: Only supported for node OS distro [gci ubuntu custom] (not debian)
... skipping 63 lines ...
Jul 18 19:51:36.569: INFO: PersistentVolumeClaim pvc-zpw8j found but phase is Pending instead of Bound.
Jul 18 19:51:38.600: INFO: PersistentVolumeClaim pvc-zpw8j found and phase=Bound (10.188583922s)
Jul 18 19:51:38.601: INFO: Waiting up to 3m0s for PersistentVolume local-fj5v6 to have phase Bound
Jul 18 19:51:38.631: INFO: PersistentVolume local-fj5v6 found and phase=Bound (30.04173ms)
STEP: Creating pod pod-subpath-test-preprovisionedpv-cxfs
STEP: Creating a pod to test atomic-volume-subpath
Jul 18 19:51:38.740: INFO: Waiting up to 5m0s for pod "pod-subpath-test-preprovisionedpv-cxfs" in namespace "provisioning-4470" to be "Succeeded or Failed"
Jul 18 19:51:38.771: INFO: Pod "pod-subpath-test-preprovisionedpv-cxfs": Phase="Pending", Reason="", readiness=false. Elapsed: 30.445395ms
Jul 18 19:51:40.802: INFO: Pod "pod-subpath-test-preprovisionedpv-cxfs": Phase="Pending", Reason="", readiness=false. Elapsed: 2.062074282s
Jul 18 19:51:42.838: INFO: Pod "pod-subpath-test-preprovisionedpv-cxfs": Phase="Pending", Reason="", readiness=false. Elapsed: 4.097826846s
Jul 18 19:51:44.871: INFO: Pod "pod-subpath-test-preprovisionedpv-cxfs": Phase="Pending", Reason="", readiness=false. Elapsed: 6.130359396s
Jul 18 19:51:46.904: INFO: Pod "pod-subpath-test-preprovisionedpv-cxfs": Phase="Pending", Reason="", readiness=false. Elapsed: 8.163507387s
Jul 18 19:51:48.936: INFO: Pod "pod-subpath-test-preprovisionedpv-cxfs": Phase="Pending", Reason="", readiness=false. Elapsed: 10.195776711s
... skipping 8 lines ...
Jul 18 19:52:07.234: INFO: Pod "pod-subpath-test-preprovisionedpv-cxfs": Phase="Running", Reason="", readiness=true. Elapsed: 28.493594453s
Jul 18 19:52:09.265: INFO: Pod "pod-subpath-test-preprovisionedpv-cxfs": Phase="Running", Reason="", readiness=true. Elapsed: 30.524712877s
Jul 18 19:52:11.303: INFO: Pod "pod-subpath-test-preprovisionedpv-cxfs": Phase="Running", Reason="", readiness=true. Elapsed: 32.562567139s
Jul 18 19:52:13.335: INFO: Pod "pod-subpath-test-preprovisionedpv-cxfs": Phase="Running", Reason="", readiness=true. Elapsed: 34.594523458s
Jul 18 19:52:15.367: INFO: Pod "pod-subpath-test-preprovisionedpv-cxfs": Phase="Succeeded", Reason="", readiness=false. Elapsed: 36.626751045s
STEP: Saw pod success
Jul 18 19:52:15.367: INFO: Pod "pod-subpath-test-preprovisionedpv-cxfs" satisfied condition "Succeeded or Failed"
Jul 18 19:52:15.398: INFO: Trying to get logs from node 10.116.0.5 pod pod-subpath-test-preprovisionedpv-cxfs container test-container-subpath-preprovisionedpv-cxfs: <nil>
STEP: delete the pod
Jul 18 19:52:15.479: INFO: Waiting for pod pod-subpath-test-preprovisionedpv-cxfs to disappear
Jul 18 19:52:15.510: INFO: Pod pod-subpath-test-preprovisionedpv-cxfs no longer exists
STEP: Deleting pod pod-subpath-test-preprovisionedpv-cxfs
Jul 18 19:52:15.510: INFO: Deleting pod "pod-subpath-test-preprovisionedpv-cxfs" in namespace "provisioning-4470"
... skipping 19 lines ...
  /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/storage/in_tree_volumes.go:58
    [Testpattern: Pre-provisioned PV (default fs)] subPath
    /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/storage/framework/testsuite.go:50
      should support file as subpath [LinuxOnly]
      /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/storage/testsuites/subpath.go:230
------------------------------
{"msg":"PASSED [sig-storage] In-tree Volumes [Driver: local][LocalVolumeType: dir-bindmounted] [Testpattern: Pre-provisioned PV (default fs)] subPath should support file as subpath [LinuxOnly]","total":-1,"completed":2,"skipped":7,"failed":0}

SS
------------------------------
[BeforeEach] [Testpattern: Inline-volume (default fs)] subPath
  /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/storage/framework/testsuite.go:51
Jul 18 19:52:16.064: INFO: Driver csi-hostpath doesn't support InlineVolume -- skipping
... skipping 123 lines ...
  /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/storage/in_tree_volumes.go:58
    [Testpattern: Pre-provisioned PV (default fs)] subPath
    /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/storage/framework/testsuite.go:50
      should be able to unmount after the subpath directory is deleted [LinuxOnly]
      /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/storage/testsuites/subpath.go:444
------------------------------
{"msg":"PASSED [sig-storage] In-tree Volumes [Driver: local][LocalVolumeType: block] [Testpattern: Pre-provisioned PV (default fs)] subPath should be able to unmount after the subpath directory is deleted [LinuxOnly]","total":-1,"completed":2,"skipped":4,"failed":0}

S
------------------------------
[BeforeEach] [Testpattern: Dynamic PV (default fs)] capacity
  /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/storage/framework/testsuite.go:51
Jul 18 19:52:17.155: INFO: Driver local doesn't support DynamicPV -- skipping
... skipping 23 lines ...
Jul 18 19:52:02.787: INFO: >>> kubeConfig: /root/.kube/config
STEP: Building a namespace api object, basename containers
STEP: Waiting for a default service account to be provisioned in namespace
[It] should be able to override the image's default command (docker entrypoint) [NodeConformance] [Conformance]
  /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:630
STEP: Creating a pod to test override command
Jul 18 19:52:02.975: INFO: Waiting up to 5m0s for pod "client-containers-a8f19104-991c-4db4-a56b-7d4445f57e84" in namespace "containers-995" to be "Succeeded or Failed"
Jul 18 19:52:03.005: INFO: Pod "client-containers-a8f19104-991c-4db4-a56b-7d4445f57e84": Phase="Pending", Reason="", readiness=false. Elapsed: 30.12722ms
Jul 18 19:52:05.040: INFO: Pod "client-containers-a8f19104-991c-4db4-a56b-7d4445f57e84": Phase="Pending", Reason="", readiness=false. Elapsed: 2.064422893s
Jul 18 19:52:07.074: INFO: Pod "client-containers-a8f19104-991c-4db4-a56b-7d4445f57e84": Phase="Pending", Reason="", readiness=false. Elapsed: 4.098576421s
Jul 18 19:52:09.109: INFO: Pod "client-containers-a8f19104-991c-4db4-a56b-7d4445f57e84": Phase="Pending", Reason="", readiness=false. Elapsed: 6.133642179s
Jul 18 19:52:11.142: INFO: Pod "client-containers-a8f19104-991c-4db4-a56b-7d4445f57e84": Phase="Pending", Reason="", readiness=false. Elapsed: 8.166394927s
Jul 18 19:52:13.173: INFO: Pod "client-containers-a8f19104-991c-4db4-a56b-7d4445f57e84": Phase="Pending", Reason="", readiness=false. Elapsed: 10.197223489s
Jul 18 19:52:15.204: INFO: Pod "client-containers-a8f19104-991c-4db4-a56b-7d4445f57e84": Phase="Pending", Reason="", readiness=false. Elapsed: 12.229167265s
Jul 18 19:52:17.256: INFO: Pod "client-containers-a8f19104-991c-4db4-a56b-7d4445f57e84": Phase="Succeeded", Reason="", readiness=false. Elapsed: 14.28047415s
STEP: Saw pod success
Jul 18 19:52:17.256: INFO: Pod "client-containers-a8f19104-991c-4db4-a56b-7d4445f57e84" satisfied condition "Succeeded or Failed"
Jul 18 19:52:17.287: INFO: Trying to get logs from node 10.116.0.5 pod client-containers-a8f19104-991c-4db4-a56b-7d4445f57e84 container agnhost-container: <nil>
STEP: delete the pod
Jul 18 19:52:17.381: INFO: Waiting for pod client-containers-a8f19104-991c-4db4-a56b-7d4445f57e84 to disappear
Jul 18 19:52:17.411: INFO: Pod client-containers-a8f19104-991c-4db4-a56b-7d4445f57e84 no longer exists
[AfterEach] [sig-node] Docker Containers
  /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:186
... skipping 4 lines ...
• [SLOW TEST:14.717 seconds]
[sig-node] Docker Containers
/workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/common/node/framework.go:23
  should be able to override the image's default command (docker entrypoint) [NodeConformance] [Conformance]
  /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:630
------------------------------
{"msg":"PASSED [sig-node] Docker Containers should be able to override the image's default command (docker entrypoint) [NodeConformance] [Conformance]","total":-1,"completed":4,"skipped":40,"failed":0}

SS
------------------------------
[BeforeEach] [sig-api-machinery] API priority and fairness
  /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:185
STEP: Creating a kubernetes client
... skipping 11 lines ...
  /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:186
Jul 18 19:52:17.802: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready
STEP: Destroying namespace "apf-1925" for this suite.

•
------------------------------
{"msg":"PASSED [sig-api-machinery] API priority and fairness should ensure that requests can be classified by adding FlowSchema and PriorityLevelConfiguration","total":-1,"completed":3,"skipped":12,"failed":0}

SSSSSSSS
------------------------------
[BeforeEach] [sig-apps] ReplicaSet
  /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:185
STEP: Creating a kubernetes client
... skipping 22 lines ...
• [SLOW TEST:26.388 seconds]
[sig-apps] ReplicaSet
/workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/apps/framework.go:23
  Replace and Patch tests [Conformance]
  /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:630
------------------------------
{"msg":"PASSED [sig-apps] ReplicaSet Replace and Patch tests [Conformance]","total":-1,"completed":3,"skipped":7,"failed":0}

S
------------------------------
[BeforeEach] [Testpattern: Inline-volume (ext4)] volumes
  /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/storage/framework/testsuite.go:51
Jul 18 19:52:17.956: INFO: Driver local doesn't support InlineVolume -- skipping
... skipping 73 lines ...
Jul 18 19:51:23.517: INFO: PersistentVolumeClaim pvc-cghzv found and phase=Bound (16.292823511s)
Jul 18 19:51:23.517: INFO: Waiting up to 3m0s for PersistentVolume nfs-gwrrs to have phase Bound
Jul 18 19:51:23.550: INFO: PersistentVolume nfs-gwrrs found and phase=Bound (32.580936ms)
STEP: Checking pod has write access to PersistentVolume
Jul 18 19:51:23.611: INFO: Creating nfs test pod
Jul 18 19:51:23.651: INFO: Pod should terminate with exitcode 0 (success)
Jul 18 19:51:23.652: INFO: Waiting up to 5m0s for pod "pvc-tester-c5mg8" in namespace "pv-6492" to be "Succeeded or Failed"
Jul 18 19:51:23.682: INFO: Pod "pvc-tester-c5mg8": Phase="Pending", Reason="", readiness=false. Elapsed: 30.778852ms
Jul 18 19:51:25.714: INFO: Pod "pvc-tester-c5mg8": Phase="Pending", Reason="", readiness=false. Elapsed: 2.062786291s
Jul 18 19:51:27.746: INFO: Pod "pvc-tester-c5mg8": Phase="Pending", Reason="", readiness=false. Elapsed: 4.09460481s
Jul 18 19:51:29.777: INFO: Pod "pvc-tester-c5mg8": Phase="Pending", Reason="", readiness=false. Elapsed: 6.125632517s
Jul 18 19:51:31.808: INFO: Pod "pvc-tester-c5mg8": Phase="Pending", Reason="", readiness=false. Elapsed: 8.156281499s
Jul 18 19:51:33.839: INFO: Pod "pvc-tester-c5mg8": Phase="Pending", Reason="", readiness=false. Elapsed: 10.187375085s
... skipping 5 lines ...
Jul 18 19:51:46.029: INFO: Pod "pvc-tester-c5mg8": Phase="Pending", Reason="", readiness=false. Elapsed: 22.377035042s
Jul 18 19:51:48.061: INFO: Pod "pvc-tester-c5mg8": Phase="Pending", Reason="", readiness=false. Elapsed: 24.409561151s
Jul 18 19:51:50.092: INFO: Pod "pvc-tester-c5mg8": Phase="Pending", Reason="", readiness=false. Elapsed: 26.440280217s
Jul 18 19:51:52.125: INFO: Pod "pvc-tester-c5mg8": Phase="Pending", Reason="", readiness=false. Elapsed: 28.473198412s
Jul 18 19:51:54.157: INFO: Pod "pvc-tester-c5mg8": Phase="Succeeded", Reason="", readiness=false. Elapsed: 30.50589043s
STEP: Saw pod success
Jul 18 19:51:54.158: INFO: Pod "pvc-tester-c5mg8" satisfied condition "Succeeded or Failed"
Jul 18 19:51:54.158: INFO: Pod pvc-tester-c5mg8 succeeded 
Jul 18 19:51:54.160: INFO: Deleting pod "pvc-tester-c5mg8" in namespace "pv-6492"
Jul 18 19:51:54.204: INFO: Wait up to 5m0s for pod "pvc-tester-c5mg8" to be fully deleted
STEP: Deleting the PVC to invoke the reclaim policy.
Jul 18 19:51:54.237: INFO: Deleting PVC pvc-cghzv to trigger reclamation of PV 
Jul 18 19:51:54.237: INFO: Deleting PersistentVolumeClaim "pvc-cghzv"
... skipping 23 lines ...
  /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/storage/persistent_volumes.go:122
    with Single PV - PVC pairs
    /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/storage/persistent_volumes.go:155
      create a PVC and non-pre-bound PV: test write access
      /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/storage/persistent_volumes.go:178
------------------------------
{"msg":"PASSED [sig-storage] PersistentVolumes NFS with Single PV - PVC pairs create a PVC and non-pre-bound PV: test write access","total":-1,"completed":2,"skipped":23,"failed":0}

S
------------------------------
[BeforeEach] [Testpattern: Pre-provisioned PV (block volmode)] volumeMode
  /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/storage/framework/testsuite.go:51
Jul 18 19:52:18.613: INFO: Only supported for providers [openstack] (not skeleton)
... skipping 78 lines ...
• [SLOW TEST:14.656 seconds]
[sig-api-machinery] AdmissionWebhook [Privileged:ClusterAdmin]
/workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/apimachinery/framework.go:23
  should deny crd creation [Conformance]
  /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:630
------------------------------
{"msg":"PASSED [sig-api-machinery] AdmissionWebhook [Privileged:ClusterAdmin] should deny crd creation [Conformance]","total":-1,"completed":4,"skipped":17,"failed":0}

S
------------------------------
[BeforeEach] [sig-network] Services
  /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:185
STEP: Creating a kubernetes client
... skipping 42 lines ...
• [SLOW TEST:21.246 seconds]
[sig-network] Services
/workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/network/common/framework.go:23
  should serve a basic endpoint from pods  [Conformance]
  /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:630
------------------------------
{"msg":"PASSED [sig-network] Services should serve a basic endpoint from pods  [Conformance]","total":-1,"completed":4,"skipped":9,"failed":0}
[BeforeEach] [Testpattern: Dynamic PV (block volmode)(allowExpansion)] volume-expand
  /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/storage/framework/testsuite.go:51
Jul 18 19:52:23.277: INFO: Only supported for providers [vsphere] (not skeleton)
[AfterEach] [Testpattern: Dynamic PV (block volmode)(allowExpansion)] volume-expand
  /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:186

... skipping 45 lines ...
  /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/storage/subpath.go:38
STEP: Setting up data
[It] should support subpaths with projected pod [LinuxOnly] [Conformance]
  /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:630
STEP: Creating pod pod-subpath-test-projected-jbzg
STEP: Creating a pod to test atomic-volume-subpath
Jul 18 19:51:50.902: INFO: Waiting up to 5m0s for pod "pod-subpath-test-projected-jbzg" in namespace "subpath-9374" to be "Succeeded or Failed"
Jul 18 19:51:50.933: INFO: Pod "pod-subpath-test-projected-jbzg": Phase="Pending", Reason="", readiness=false. Elapsed: 30.040685ms
Jul 18 19:51:52.965: INFO: Pod "pod-subpath-test-projected-jbzg": Phase="Pending", Reason="", readiness=false. Elapsed: 2.062329035s
Jul 18 19:51:54.998: INFO: Pod "pod-subpath-test-projected-jbzg": Phase="Pending", Reason="", readiness=false. Elapsed: 4.095522349s
Jul 18 19:51:57.030: INFO: Pod "pod-subpath-test-projected-jbzg": Phase="Pending", Reason="", readiness=false. Elapsed: 6.127508657s
Jul 18 19:51:59.063: INFO: Pod "pod-subpath-test-projected-jbzg": Phase="Pending", Reason="", readiness=false. Elapsed: 8.159998106s
Jul 18 19:52:01.094: INFO: Pod "pod-subpath-test-projected-jbzg": Phase="Pending", Reason="", readiness=false. Elapsed: 10.191304765s
... skipping 6 lines ...
Jul 18 19:52:15.318: INFO: Pod "pod-subpath-test-projected-jbzg": Phase="Running", Reason="", readiness=true. Elapsed: 24.415382653s
Jul 18 19:52:17.349: INFO: Pod "pod-subpath-test-projected-jbzg": Phase="Running", Reason="", readiness=true. Elapsed: 26.446693057s
Jul 18 19:52:19.380: INFO: Pod "pod-subpath-test-projected-jbzg": Phase="Running", Reason="", readiness=true. Elapsed: 28.477379227s
Jul 18 19:52:21.412: INFO: Pod "pod-subpath-test-projected-jbzg": Phase="Running", Reason="", readiness=true. Elapsed: 30.509183565s
Jul 18 19:52:23.443: INFO: Pod "pod-subpath-test-projected-jbzg": Phase="Succeeded", Reason="", readiness=false. Elapsed: 32.54067471s
STEP: Saw pod success
Jul 18 19:52:23.443: INFO: Pod "pod-subpath-test-projected-jbzg" satisfied condition "Succeeded or Failed"
Jul 18 19:52:23.481: INFO: Trying to get logs from node 10.116.0.7 pod pod-subpath-test-projected-jbzg container test-container-subpath-projected-jbzg: <nil>
STEP: delete the pod
Jul 18 19:52:23.553: INFO: Waiting for pod pod-subpath-test-projected-jbzg to disappear
Jul 18 19:52:23.587: INFO: Pod pod-subpath-test-projected-jbzg no longer exists
STEP: Deleting pod pod-subpath-test-projected-jbzg
Jul 18 19:52:23.587: INFO: Deleting pod "pod-subpath-test-projected-jbzg" in namespace "subpath-9374"
... skipping 8 lines ...
/workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/storage/utils/framework.go:23
  Atomic writer volumes
  /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/storage/subpath.go:34
    should support subpaths with projected pod [LinuxOnly] [Conformance]
    /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:630
------------------------------
{"msg":"PASSED [sig-storage] Subpath Atomic writer volumes should support subpaths with projected pod [LinuxOnly] [Conformance]","total":-1,"completed":3,"skipped":9,"failed":0}

S
------------------------------
[BeforeEach] [sig-storage] PersistentVolumes-local 
  /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:185
STEP: Creating a kubernetes client
... skipping 57 lines ...
  /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/storage/persistent_volumes-local.go:192
    One pod requesting one prebound PVC
    /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/storage/persistent_volumes-local.go:209
      should be able to mount volume and write from pod1
      /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/storage/persistent_volumes-local.go:238
------------------------------
{"msg":"PASSED [sig-storage] PersistentVolumes-local  [Volume type: dir] One pod requesting one prebound PVC should be able to mount volume and write from pod1","total":-1,"completed":4,"skipped":12,"failed":0}

S
------------------------------
[BeforeEach] [sig-storage] Ephemeralstorage
  /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:185
STEP: Creating a kubernetes client
... skipping 18 lines ...
/workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/storage/utils/framework.go:23
  When pod refers to non-existent ephemeral storage
  /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/storage/ephemeral_volume.go:53
    should allow deletion of pod with invalid volume : secret
    /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/storage/ephemeral_volume.go:55
------------------------------
{"msg":"PASSED [sig-storage] Ephemeralstorage When pod refers to non-existent ephemeral storage should allow deletion of pod with invalid volume : secret","total":-1,"completed":5,"skipped":84,"failed":0}
[BeforeEach] [Testpattern: Dynamic PV (immediate binding)] topology
  /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/storage/framework/testsuite.go:51
Jul 18 19:52:24.022: INFO: Driver emptydir doesn't support DynamicPV -- skipping
[AfterEach] [Testpattern: Dynamic PV (immediate binding)] topology
  /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:186

... skipping 22 lines ...
Jul 18 19:50:54.595: INFO: >>> kubeConfig: /root/.kube/config
STEP: Building a namespace api object, basename volume
STEP: Waiting for a default service account to be provisioned in namespace
[It] should store data
  /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/storage/testsuites/volumes.go:159
Jul 18 19:50:54.748: INFO: In-tree plugin kubernetes.io/host-path is not migrated, not validating any metrics
Jul 18 19:50:54.816: INFO: Waiting up to 5m0s for pod "hostpath-symlink-prep-volume-1020" in namespace "volume-1020" to be "Succeeded or Failed"
Jul 18 19:50:54.846: INFO: Pod "hostpath-symlink-prep-volume-1020": Phase="Pending", Reason="", readiness=false. Elapsed: 30.541212ms
Jul 18 19:50:56.878: INFO: Pod "hostpath-symlink-prep-volume-1020": Phase="Pending", Reason="", readiness=false. Elapsed: 2.0625623s
Jul 18 19:50:58.914: INFO: Pod "hostpath-symlink-prep-volume-1020": Phase="Pending", Reason="", readiness=false. Elapsed: 4.098016636s
Jul 18 19:51:00.945: INFO: Pod "hostpath-symlink-prep-volume-1020": Phase="Pending", Reason="", readiness=false. Elapsed: 6.129349667s
Jul 18 19:51:03.118: INFO: Pod "hostpath-symlink-prep-volume-1020": Phase="Succeeded", Reason="", readiness=false. Elapsed: 8.302593615s
STEP: Saw pod success
Jul 18 19:51:03.118: INFO: Pod "hostpath-symlink-prep-volume-1020" satisfied condition "Succeeded or Failed"
Jul 18 19:51:03.118: INFO: Deleting pod "hostpath-symlink-prep-volume-1020" in namespace "volume-1020"
Jul 18 19:51:03.162: INFO: Wait up to 5m0s for pod "hostpath-symlink-prep-volume-1020" to be fully deleted
Jul 18 19:51:03.196: INFO: Creating resource for inline volume
STEP: starting hostpathsymlink-injector
STEP: Writing text file contents in the container.
Jul 18 19:51:11.299: INFO: Running '/tmp/kubectl2220469345/kubectl --server=https://143.244.215.182 --kubeconfig=/root/.kube/config --namespace=volume-1020 exec hostpathsymlink-injector --namespace=volume-1020 -- /bin/sh -c echo 'Hello from hostPathSymlink from namespace volume-1020' > /opt/0/index.html'
... skipping 54 lines ...
Jul 18 19:51:59.858: INFO: Pod hostpathsymlink-client still exists
Jul 18 19:52:01.828: INFO: Waiting for pod hostpathsymlink-client to disappear
Jul 18 19:52:01.861: INFO: Pod hostpathsymlink-client still exists
Jul 18 19:52:03.827: INFO: Waiting for pod hostpathsymlink-client to disappear
Jul 18 19:52:03.858: INFO: Pod hostpathsymlink-client no longer exists
STEP: cleaning the environment after hostpathsymlink
Jul 18 19:52:03.902: INFO: Waiting up to 5m0s for pod "hostpath-symlink-prep-volume-1020" in namespace "volume-1020" to be "Succeeded or Failed"
Jul 18 19:52:03.936: INFO: Pod "hostpath-symlink-prep-volume-1020": Phase="Pending", Reason="", readiness=false. Elapsed: 33.984139ms
Jul 18 19:52:05.969: INFO: Pod "hostpath-symlink-prep-volume-1020": Phase="Pending", Reason="", readiness=false. Elapsed: 2.066864871s
Jul 18 19:52:08.002: INFO: Pod "hostpath-symlink-prep-volume-1020": Phase="Pending", Reason="", readiness=false. Elapsed: 4.1007705s
Jul 18 19:52:10.034: INFO: Pod "hostpath-symlink-prep-volume-1020": Phase="Pending", Reason="", readiness=false. Elapsed: 6.132349672s
Jul 18 19:52:12.066: INFO: Pod "hostpath-symlink-prep-volume-1020": Phase="Pending", Reason="", readiness=false. Elapsed: 8.164176032s
Jul 18 19:52:14.097: INFO: Pod "hostpath-symlink-prep-volume-1020": Phase="Pending", Reason="", readiness=false. Elapsed: 10.195308188s
Jul 18 19:52:16.128: INFO: Pod "hostpath-symlink-prep-volume-1020": Phase="Pending", Reason="", readiness=false. Elapsed: 12.226139444s
Jul 18 19:52:18.161: INFO: Pod "hostpath-symlink-prep-volume-1020": Phase="Pending", Reason="", readiness=false. Elapsed: 14.259109446s
Jul 18 19:52:20.192: INFO: Pod "hostpath-symlink-prep-volume-1020": Phase="Pending", Reason="", readiness=false. Elapsed: 16.290136999s
Jul 18 19:52:22.224: INFO: Pod "hostpath-symlink-prep-volume-1020": Phase="Pending", Reason="", readiness=false. Elapsed: 18.322574803s
Jul 18 19:52:24.257: INFO: Pod "hostpath-symlink-prep-volume-1020": Phase="Succeeded", Reason="", readiness=false. Elapsed: 20.355630512s
STEP: Saw pod success
Jul 18 19:52:24.257: INFO: Pod "hostpath-symlink-prep-volume-1020" satisfied condition "Succeeded or Failed"
Jul 18 19:52:24.257: INFO: Deleting pod "hostpath-symlink-prep-volume-1020" in namespace "volume-1020"
Jul 18 19:52:24.302: INFO: Wait up to 5m0s for pod "hostpath-symlink-prep-volume-1020" to be fully deleted
[AfterEach] [Testpattern: Inline-volume (default fs)] volumes
  /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:186
Jul 18 19:52:24.334: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready
STEP: Destroying namespace "volume-1020" for this suite.
... skipping 6 lines ...
  /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/storage/in_tree_volumes.go:58
    [Testpattern: Inline-volume (default fs)] volumes
    /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/storage/framework/testsuite.go:50
      should store data
      /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/storage/testsuites/volumes.go:159
------------------------------
{"msg":"PASSED [sig-storage] In-tree Volumes [Driver: hostPathSymlink] [Testpattern: Inline-volume (default fs)] volumes should store data","total":-1,"completed":2,"skipped":4,"failed":0}

S
------------------------------
[BeforeEach] [Testpattern: Dynamic PV (default fs)] fsgroupchangepolicy
  /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/storage/framework/testsuite.go:51
Jul 18 19:52:24.463: INFO: Driver hostPathSymlink doesn't support DynamicPV -- skipping
... skipping 57 lines ...
• [SLOW TEST:25.732 seconds]
[sig-node] PreStop
/workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/node/framework.go:23
  should call prestop when killing a pod  [Conformance]
  /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:630
------------------------------
{"msg":"PASSED [sig-node] PreStop should call prestop when killing a pod  [Conformance]","total":-1,"completed":7,"skipped":50,"failed":0}

SSSSS
------------------------------
[BeforeEach] [Testpattern: Dynamic PV (block volmode)] volumeMode
  /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/storage/framework/testsuite.go:51
Jul 18 19:52:27.089: INFO: Driver local doesn't support DynamicPV -- skipping
... skipping 35 lines ...
/workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/apimachinery/framework.go:23
  Simple CustomResourceDefinition
  /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/apimachinery/custom_resource_definition.go:48
    creating/deleting custom resource definition objects works  [Conformance]
    /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:630
------------------------------
{"msg":"PASSED [sig-api-machinery] CustomResourceDefinition resources [Privileged:ClusterAdmin] Simple CustomResourceDefinition creating/deleting custom resource definition objects works  [Conformance]","total":-1,"completed":5,"skipped":13,"failed":0}

SSS
------------------------------
[BeforeEach] [sig-instrumentation] MetricsGrabber
  /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:185
STEP: Creating a kubernetes client
... skipping 11 lines ...
  /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:186
Jul 18 19:52:29.495: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready
STEP: Destroying namespace "metrics-grabber-5765" for this suite.

•
------------------------------
{"msg":"PASSED [sig-instrumentation] MetricsGrabber should grab all metrics from a ControllerManager.","total":-1,"completed":6,"skipped":16,"failed":0}

SSSS
------------------------------
[BeforeEach] [Testpattern: Inline-volume (default fs)] subPath
  /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/storage/framework/testsuite.go:51
Jul 18 19:52:29.586: INFO: Driver local doesn't support InlineVolume -- skipping
... skipping 49 lines ...
[sig-storage] In-tree Volumes
/workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/storage/utils/framework.go:23
  [Driver: local][LocalVolumeType: tmpfs]
  /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/storage/in_tree_volumes.go:58
    [Testpattern: Dynamic PV (delayed binding)] topology
    /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/storage/framework/testsuite.go:50
      should fail to schedule a pod which has topologies that conflict with AllowedTopologies [BeforeEach]
      /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/storage/testsuites/topology.go:192

      Driver local doesn't support DynamicPV -- skipping

      /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/storage/framework/testsuite.go:116
------------------------------
... skipping 16 lines ...
  /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:186
Jul 18 19:52:29.941: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready
STEP: Destroying namespace "svcaccounts-7125" for this suite.

•
------------------------------
{"msg":"PASSED [sig-auth] ServiceAccounts should run through the lifecycle of a ServiceAccount [Conformance]","total":-1,"completed":7,"skipped":25,"failed":0}

SSSS
------------------------------
[BeforeEach] [Testpattern: Pre-provisioned PV (default fs)] subPath
  /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/storage/framework/testsuite.go:51
[BeforeEach] [Testpattern: Pre-provisioned PV (default fs)] subPath
... skipping 24 lines ...
Jul 18 19:52:20.562: INFO: PersistentVolumeClaim pvc-c7259 found but phase is Pending instead of Bound.
Jul 18 19:52:22.594: INFO: PersistentVolumeClaim pvc-c7259 found and phase=Bound (10.193165173s)
Jul 18 19:52:22.594: INFO: Waiting up to 3m0s for PersistentVolume local-cj58z to have phase Bound
Jul 18 19:52:22.625: INFO: PersistentVolume local-cj58z found and phase=Bound (31.000924ms)
STEP: Creating pod pod-subpath-test-preprovisionedpv-gkvd
STEP: Creating a pod to test subpath
Jul 18 19:52:22.724: INFO: Waiting up to 5m0s for pod "pod-subpath-test-preprovisionedpv-gkvd" in namespace "provisioning-2156" to be "Succeeded or Failed"
Jul 18 19:52:22.756: INFO: Pod "pod-subpath-test-preprovisionedpv-gkvd": Phase="Pending", Reason="", readiness=false. Elapsed: 31.290973ms
Jul 18 19:52:24.789: INFO: Pod "pod-subpath-test-preprovisionedpv-gkvd": Phase="Pending", Reason="", readiness=false. Elapsed: 2.064828433s
Jul 18 19:52:26.825: INFO: Pod "pod-subpath-test-preprovisionedpv-gkvd": Phase="Pending", Reason="", readiness=false. Elapsed: 4.100452016s
Jul 18 19:52:28.857: INFO: Pod "pod-subpath-test-preprovisionedpv-gkvd": Phase="Succeeded", Reason="", readiness=false. Elapsed: 6.133072959s
STEP: Saw pod success
Jul 18 19:52:28.857: INFO: Pod "pod-subpath-test-preprovisionedpv-gkvd" satisfied condition "Succeeded or Failed"
Jul 18 19:52:28.891: INFO: Trying to get logs from node 10.116.0.7 pod pod-subpath-test-preprovisionedpv-gkvd container test-container-subpath-preprovisionedpv-gkvd: <nil>
STEP: delete the pod
Jul 18 19:52:28.970: INFO: Waiting for pod pod-subpath-test-preprovisionedpv-gkvd to disappear
Jul 18 19:52:29.002: INFO: Pod pod-subpath-test-preprovisionedpv-gkvd no longer exists
STEP: Deleting pod pod-subpath-test-preprovisionedpv-gkvd
Jul 18 19:52:29.002: INFO: Deleting pod "pod-subpath-test-preprovisionedpv-gkvd" in namespace "provisioning-2156"
... skipping 26 lines ...
  /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/storage/in_tree_volumes.go:58
    [Testpattern: Pre-provisioned PV (default fs)] subPath
    /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/storage/framework/testsuite.go:50
      should support existing single file [LinuxOnly]
      /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/storage/testsuites/subpath.go:219
------------------------------
{"msg":"PASSED [sig-storage] In-tree Volumes [Driver: local][LocalVolumeType: blockfs] [Testpattern: Pre-provisioned PV (default fs)] subPath should support existing single file [LinuxOnly]","total":-1,"completed":5,"skipped":24,"failed":0}
[BeforeEach] [Testpattern: Dynamic PV (default fs)] volumes
  /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/storage/framework/testsuite.go:51
Jul 18 19:52:30.402: INFO: Only supported for providers [azure] (not skeleton)
[AfterEach] [Testpattern: Dynamic PV (default fs)] volumes
  /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:186

... skipping 58 lines ...
• [SLOW TEST:14.422 seconds]
[sig-apps] DisruptionController
/workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/apps/framework.go:23
  evictions: no PDB => should allow an eviction
  /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/apps/disruption.go:267
------------------------------
{"msg":"PASSED [sig-apps] DisruptionController evictions: no PDB =\u003e should allow an eviction","total":-1,"completed":5,"skipped":42,"failed":0}

SS
------------------------------
[BeforeEach] [sig-storage] EmptyDir volumes
  /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:185
STEP: Creating a kubernetes client
Jul 18 19:52:17.906: INFO: >>> kubeConfig: /root/.kube/config
STEP: Building a namespace api object, basename emptydir
STEP: Waiting for a default service account to be provisioned in namespace
[It] should support (non-root,0777,default) [LinuxOnly] [NodeConformance] [Conformance]
  /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:630
STEP: Creating a pod to test emptydir 0777 on node default medium
Jul 18 19:52:18.106: INFO: Waiting up to 5m0s for pod "pod-89e7c461-8e38-4767-8f04-053978429209" in namespace "emptydir-1282" to be "Succeeded or Failed"
Jul 18 19:52:18.138: INFO: Pod "pod-89e7c461-8e38-4767-8f04-053978429209": Phase="Pending", Reason="", readiness=false. Elapsed: 31.544436ms
Jul 18 19:52:20.171: INFO: Pod "pod-89e7c461-8e38-4767-8f04-053978429209": Phase="Pending", Reason="", readiness=false. Elapsed: 2.064396885s
Jul 18 19:52:22.210: INFO: Pod "pod-89e7c461-8e38-4767-8f04-053978429209": Phase="Pending", Reason="", readiness=false. Elapsed: 4.103420587s
Jul 18 19:52:24.242: INFO: Pod "pod-89e7c461-8e38-4767-8f04-053978429209": Phase="Pending", Reason="", readiness=false. Elapsed: 6.135879819s
Jul 18 19:52:26.274: INFO: Pod "pod-89e7c461-8e38-4767-8f04-053978429209": Phase="Pending", Reason="", readiness=false. Elapsed: 8.16736257s
Jul 18 19:52:28.309: INFO: Pod "pod-89e7c461-8e38-4767-8f04-053978429209": Phase="Pending", Reason="", readiness=false. Elapsed: 10.202232829s
Jul 18 19:52:30.342: INFO: Pod "pod-89e7c461-8e38-4767-8f04-053978429209": Phase="Pending", Reason="", readiness=false. Elapsed: 12.235388344s
Jul 18 19:52:32.374: INFO: Pod "pod-89e7c461-8e38-4767-8f04-053978429209": Phase="Succeeded", Reason="", readiness=false. Elapsed: 14.267894439s
STEP: Saw pod success
Jul 18 19:52:32.374: INFO: Pod "pod-89e7c461-8e38-4767-8f04-053978429209" satisfied condition "Succeeded or Failed"
Jul 18 19:52:32.407: INFO: Trying to get logs from node 10.116.0.5 pod pod-89e7c461-8e38-4767-8f04-053978429209 container test-container: <nil>
STEP: delete the pod
Jul 18 19:52:32.491: INFO: Waiting for pod pod-89e7c461-8e38-4767-8f04-053978429209 to disappear
Jul 18 19:52:32.522: INFO: Pod pod-89e7c461-8e38-4767-8f04-053978429209 no longer exists
[AfterEach] [sig-storage] EmptyDir volumes
  /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:186
... skipping 4 lines ...
• [SLOW TEST:14.714 seconds]
[sig-storage] EmptyDir volumes
/workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/common/storage/framework.go:23
  should support (non-root,0777,default) [LinuxOnly] [NodeConformance] [Conformance]
  /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:630
------------------------------
{"msg":"PASSED [sig-storage] EmptyDir volumes should support (non-root,0777,default) [LinuxOnly] [NodeConformance] [Conformance]","total":-1,"completed":4,"skipped":20,"failed":0}

S
------------------------------
[BeforeEach] [sig-api-machinery] AdmissionWebhook [Privileged:ClusterAdmin]
  /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:185
STEP: Creating a kubernetes client
... skipping 10 lines ...
Jul 18 19:52:26.969: INFO: deployment status: v1.DeploymentStatus{ObservedGeneration:1, Replicas:1, UpdatedReplicas:1, ReadyReplicas:0, AvailableReplicas:0, UnavailableReplicas:1, Conditions:[]v1.DeploymentCondition{v1.DeploymentCondition{Type:"Available", Status:"False", LastUpdateTime:v1.Time{Time:time.Time{wall:0x0, ext:63762234744, loc:(*time.Location)(0x9ddf5a0)}}, LastTransitionTime:v1.Time{Time:time.Time{wall:0x0, ext:63762234744, loc:(*time.Location)(0x9ddf5a0)}}, Reason:"MinimumReplicasUnavailable", Message:"Deployment does not have minimum availability."}, v1.DeploymentCondition{Type:"Progressing", Status:"True", LastUpdateTime:v1.Time{Time:time.Time{wall:0x0, ext:63762234744, loc:(*time.Location)(0x9ddf5a0)}}, LastTransitionTime:v1.Time{Time:time.Time{wall:0x0, ext:63762234744, loc:(*time.Location)(0x9ddf5a0)}}, Reason:"ReplicaSetUpdated", Message:"ReplicaSet \"sample-webhook-deployment-78988fc6cd\" is progressing."}}, CollisionCount:(*int32)(nil)}
Jul 18 19:52:28.970: INFO: deployment status: v1.DeploymentStatus{ObservedGeneration:1, Replicas:1, UpdatedReplicas:1, ReadyReplicas:0, AvailableReplicas:0, UnavailableReplicas:1, Conditions:[]v1.DeploymentCondition{v1.DeploymentCondition{Type:"Available", Status:"False", LastUpdateTime:v1.Time{Time:time.Time{wall:0x0, ext:63762234744, loc:(*time.Location)(0x9ddf5a0)}}, LastTransitionTime:v1.Time{Time:time.Time{wall:0x0, ext:63762234744, loc:(*time.Location)(0x9ddf5a0)}}, Reason:"MinimumReplicasUnavailable", Message:"Deployment does not have minimum availability."}, v1.DeploymentCondition{Type:"Progressing", Status:"True", LastUpdateTime:v1.Time{Time:time.Time{wall:0x0, ext:63762234744, loc:(*time.Location)(0x9ddf5a0)}}, LastTransitionTime:v1.Time{Time:time.Time{wall:0x0, ext:63762234744, loc:(*time.Location)(0x9ddf5a0)}}, Reason:"ReplicaSetUpdated", Message:"ReplicaSet \"sample-webhook-deployment-78988fc6cd\" is progressing."}}, CollisionCount:(*int32)(nil)}
Jul 18 19:52:30.969: INFO: deployment status: v1.DeploymentStatus{ObservedGeneration:1, Replicas:1, UpdatedReplicas:1, ReadyReplicas:0, AvailableReplicas:0, UnavailableReplicas:1, Conditions:[]v1.DeploymentCondition{v1.DeploymentCondition{Type:"Available", Status:"False", LastUpdateTime:v1.Time{Time:time.Time{wall:0x0, ext:63762234744, loc:(*time.Location)(0x9ddf5a0)}}, LastTransitionTime:v1.Time{Time:time.Time{wall:0x0, ext:63762234744, loc:(*time.Location)(0x9ddf5a0)}}, Reason:"MinimumReplicasUnavailable", Message:"Deployment does not have minimum availability."}, v1.DeploymentCondition{Type:"Progressing", Status:"True", LastUpdateTime:v1.Time{Time:time.Time{wall:0x0, ext:63762234744, loc:(*time.Location)(0x9ddf5a0)}}, LastTransitionTime:v1.Time{Time:time.Time{wall:0x0, ext:63762234744, loc:(*time.Location)(0x9ddf5a0)}}, Reason:"ReplicaSetUpdated", Message:"ReplicaSet \"sample-webhook-deployment-78988fc6cd\" is progressing."}}, CollisionCount:(*int32)(nil)}
STEP: Deploying the webhook service
STEP: Verifying the service has paired with the endpoint
Jul 18 19:52:34.014: INFO: Waiting for amount of service:e2e-test-webhook endpoints to be 1
[It] should unconditionally reject operations on fail closed webhook [Conformance]
  /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:630
STEP: Registering a webhook that server cannot talk to, with fail closed policy, via the AdmissionRegistration API
STEP: create a namespace for the webhook
STEP: create a configmap should be unconditionally rejected by the webhook
[AfterEach] [sig-api-machinery] AdmissionWebhook [Privileged:ClusterAdmin]
  /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:186
Jul 18 19:52:34.280: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready
STEP: Destroying namespace "webhook-4829" for this suite.
... skipping 2 lines ...
  /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/apimachinery/webhook.go:102


• [SLOW TEST:10.650 seconds]
[sig-api-machinery] AdmissionWebhook [Privileged:ClusterAdmin]
/workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/apimachinery/framework.go:23
  should unconditionally reject operations on fail closed webhook [Conformance]
  /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:630
------------------------------
[BeforeEach] [sig-network] HostPort
  /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:185
STEP: Creating a kubernetes client
Jul 18 19:51:31.804: INFO: >>> kubeConfig: /root/.kube/config
... skipping 57 lines ...
• [SLOW TEST:64.516 seconds]
[sig-network] HostPort
/workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/network/common/framework.go:23
  validates that there is no conflict between pods with same hostPort but different hostIP and protocol [LinuxOnly] [Conformance]
  /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:630
------------------------------
{"msg":"PASSED [sig-network] HostPort validates that there is no conflict between pods with same hostPort but different hostIP and protocol [LinuxOnly] [Conformance]","total":-1,"completed":2,"skipped":4,"failed":0}

SSSSSS
------------------------------
[BeforeEach] [Testpattern: Inline-volume (default fs)] subPath
  /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/storage/framework/testsuite.go:51
Jul 18 19:52:36.359: INFO: Only supported for providers [azure] (not skeleton)
... skipping 202 lines ...
  /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/storage/persistent_volumes-local.go:192
    Two pods mounting a local volume at the same time
    /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/storage/persistent_volumes-local.go:248
      should be able to write from pod1 and read from pod2
      /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/storage/persistent_volumes-local.go:249
------------------------------
{"msg":"PASSED [sig-storage] PersistentVolumes-local  [Volume type: dir-bindmounted] Two pods mounting a local volume at the same time should be able to write from pod1 and read from pod2","total":-1,"completed":4,"skipped":36,"failed":0}
[BeforeEach] [Testpattern: Inline-volume (ext4)] volumes
  /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/storage/framework/testsuite.go:51
Jul 18 19:52:36.739: INFO: Driver local doesn't support InlineVolume -- skipping
[AfterEach] [Testpattern: Inline-volume (ext4)] volumes
  /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:186

... skipping 184 lines ...
  /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/storage/csi_volumes.go:40
    [Testpattern: Generic Ephemeral-volume (default fs) (late-binding)] ephemeral
    /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/storage/framework/testsuite.go:50
      should support multiple inline ephemeral volumes
      /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/storage/testsuites/ephemeral.go:211
------------------------------
{"msg":"PASSED [sig-storage] CSI Volumes [Driver: csi-hostpath] [Testpattern: Generic Ephemeral-volume (default fs) (late-binding)] ephemeral should support multiple inline ephemeral volumes","total":-1,"completed":1,"skipped":22,"failed":0}
[BeforeEach] [Testpattern: Dynamic PV (ext3)] volumes
  /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/storage/framework/testsuite.go:51
Jul 18 19:52:38.377: INFO: Only supported for providers [openstack] (not skeleton)
[AfterEach] [Testpattern: Dynamic PV (ext3)] volumes
  /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:186

... skipping 25 lines ...
Jul 18 19:52:23.964: INFO: The status of Pod server-envvars-f9f2983e-f146-4497-8ab9-afd9d16e037e is Pending, waiting for it to be Running (with Ready = true)
Jul 18 19:52:25.994: INFO: The status of Pod server-envvars-f9f2983e-f146-4497-8ab9-afd9d16e037e is Pending, waiting for it to be Running (with Ready = true)
Jul 18 19:52:27.996: INFO: The status of Pod server-envvars-f9f2983e-f146-4497-8ab9-afd9d16e037e is Pending, waiting for it to be Running (with Ready = true)
Jul 18 19:52:29.995: INFO: The status of Pod server-envvars-f9f2983e-f146-4497-8ab9-afd9d16e037e is Pending, waiting for it to be Running (with Ready = true)
Jul 18 19:52:31.995: INFO: The status of Pod server-envvars-f9f2983e-f146-4497-8ab9-afd9d16e037e is Pending, waiting for it to be Running (with Ready = true)
Jul 18 19:52:33.996: INFO: The status of Pod server-envvars-f9f2983e-f146-4497-8ab9-afd9d16e037e is Running (Ready = true)
Jul 18 19:52:34.105: INFO: Waiting up to 5m0s for pod "client-envvars-c72f9e38-5dc5-4b64-8513-204981e0807a" in namespace "pods-4989" to be "Succeeded or Failed"
Jul 18 19:52:34.137: INFO: Pod "client-envvars-c72f9e38-5dc5-4b64-8513-204981e0807a": Phase="Pending", Reason="", readiness=false. Elapsed: 31.275088ms
Jul 18 19:52:36.168: INFO: Pod "client-envvars-c72f9e38-5dc5-4b64-8513-204981e0807a": Phase="Pending", Reason="", readiness=false. Elapsed: 2.062339744s
Jul 18 19:52:38.199: INFO: Pod "client-envvars-c72f9e38-5dc5-4b64-8513-204981e0807a": Phase="Succeeded", Reason="", readiness=false. Elapsed: 4.093208743s
STEP: Saw pod success
Jul 18 19:52:38.199: INFO: Pod "client-envvars-c72f9e38-5dc5-4b64-8513-204981e0807a" satisfied condition "Succeeded or Failed"
Jul 18 19:52:38.239: INFO: Trying to get logs from node 10.116.0.5 pod client-envvars-c72f9e38-5dc5-4b64-8513-204981e0807a container env3cont: <nil>
STEP: delete the pod
Jul 18 19:52:38.332: INFO: Waiting for pod client-envvars-c72f9e38-5dc5-4b64-8513-204981e0807a to disappear
Jul 18 19:52:38.362: INFO: Pod client-envvars-c72f9e38-5dc5-4b64-8513-204981e0807a no longer exists
[AfterEach] [sig-node] Pods
  /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:186
... skipping 4 lines ...
• [SLOW TEST:14.733 seconds]
[sig-node] Pods
/workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/common/node/framework.go:23
  should contain environment variables for services [NodeConformance] [Conformance]
  /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:630
------------------------------
{"msg":"PASSED [sig-node] Pods should contain environment variables for services [NodeConformance] [Conformance]","total":-1,"completed":4,"skipped":10,"failed":0}

SS
------------------------------
[BeforeEach] [Testpattern: Inline-volume (default fs)] subPath
  /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/storage/framework/testsuite.go:51
Jul 18 19:52:38.479: INFO: Driver local doesn't support InlineVolume -- skipping
... skipping 79 lines ...
  /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:186
Jul 18 19:52:38.623: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready
STEP: Destroying namespace "kubelet-test-4542" for this suite.

•
------------------------------
{"msg":"PASSED [sig-node] Kubelet when scheduling a busybox command that always fails in a pod should be possible to delete [NodeConformance] [Conformance]","total":-1,"completed":2,"skipped":23,"failed":0}

S
------------------------------
[BeforeEach] [Testpattern: Dynamic PV (immediate binding)] topology
  /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/storage/framework/testsuite.go:51
Jul 18 19:52:38.705: INFO: Driver local doesn't support DynamicPV -- skipping
... skipping 87 lines ...
[BeforeEach] [sig-apps] Job
  /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:185
STEP: Creating a kubernetes client
Jul 18 19:52:36.495: INFO: >>> kubeConfig: /root/.kube/config
STEP: Building a namespace api object, basename job
STEP: Waiting for a default service account to be provisioned in namespace
[It] should fail when exceeds active deadline
  /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/apps/job.go:253
STEP: Creating a job
STEP: Ensuring job past active deadline
[AfterEach] [sig-apps] Job
  /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:186
Jul 18 19:52:38.723: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready
STEP: Destroying namespace "job-7198" for this suite.

•
------------------------------
{"msg":"PASSED [sig-apps] Job should fail when exceeds active deadline","total":-1,"completed":3,"skipped":42,"failed":0}

SSSS
------------------------------
[BeforeEach] [sig-network] Netpol API
  /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:185
STEP: Creating a kubernetes client
... skipping 54 lines ...
• [SLOW TEST:21.534 seconds]
[sig-node] Pods
/workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/common/node/framework.go:23
  should be submitted and removed [NodeConformance] [Conformance]
  /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:630
------------------------------
{"msg":"PASSED [sig-node] Pods should be submitted and removed [NodeConformance] [Conformance]","total":-1,"completed":3,"skipped":28,"failed":0}

S
------------------------------
[BeforeEach] [sig-storage] EmptyDir volumes
  /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:185
STEP: Creating a kubernetes client
Jul 18 19:52:30.441: INFO: >>> kubeConfig: /root/.kube/config
STEP: Building a namespace api object, basename emptydir
STEP: Waiting for a default service account to be provisioned in namespace
[It] should support (root,0666,default) [LinuxOnly] [NodeConformance] [Conformance]
  /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:630
STEP: Creating a pod to test emptydir 0666 on node default medium
Jul 18 19:52:30.643: INFO: Waiting up to 5m0s for pod "pod-01d4d97b-1b71-416f-969c-074a4990e1e5" in namespace "emptydir-50" to be "Succeeded or Failed"
Jul 18 19:52:30.675: INFO: Pod "pod-01d4d97b-1b71-416f-969c-074a4990e1e5": Phase="Pending", Reason="", readiness=false. Elapsed: 31.717067ms
Jul 18 19:52:32.706: INFO: Pod "pod-01d4d97b-1b71-416f-969c-074a4990e1e5": Phase="Pending", Reason="", readiness=false. Elapsed: 2.063468163s
Jul 18 19:52:34.738: INFO: Pod "pod-01d4d97b-1b71-416f-969c-074a4990e1e5": Phase="Pending", Reason="", readiness=false. Elapsed: 4.095106058s
Jul 18 19:52:36.771: INFO: Pod "pod-01d4d97b-1b71-416f-969c-074a4990e1e5": Phase="Pending", Reason="", readiness=false. Elapsed: 6.127872335s
Jul 18 19:52:38.809: INFO: Pod "pod-01d4d97b-1b71-416f-969c-074a4990e1e5": Phase="Pending", Reason="", readiness=false. Elapsed: 8.165819613s
Jul 18 19:52:40.841: INFO: Pod "pod-01d4d97b-1b71-416f-969c-074a4990e1e5": Phase="Succeeded", Reason="", readiness=false. Elapsed: 10.197890346s
STEP: Saw pod success
Jul 18 19:52:40.841: INFO: Pod "pod-01d4d97b-1b71-416f-969c-074a4990e1e5" satisfied condition "Succeeded or Failed"
Jul 18 19:52:40.872: INFO: Trying to get logs from node 10.116.0.7 pod pod-01d4d97b-1b71-416f-969c-074a4990e1e5 container test-container: <nil>
STEP: delete the pod
Jul 18 19:52:40.948: INFO: Waiting for pod pod-01d4d97b-1b71-416f-969c-074a4990e1e5 to disappear
Jul 18 19:52:40.979: INFO: Pod pod-01d4d97b-1b71-416f-969c-074a4990e1e5 no longer exists
[AfterEach] [sig-storage] EmptyDir volumes
  /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:186
... skipping 4 lines ...
• [SLOW TEST:10.633 seconds]
[sig-storage] EmptyDir volumes
/workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/common/storage/framework.go:23
  should support (root,0666,default) [LinuxOnly] [NodeConformance] [Conformance]
  /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:630
------------------------------
{"msg":"PASSED [sig-storage] EmptyDir volumes should support (root,0666,default) [LinuxOnly] [NodeConformance] [Conformance]","total":-1,"completed":6,"skipped":32,"failed":0}

SSSSSS
------------------------------
[BeforeEach] [sig-storage] Projected downwardAPI
  /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:185
STEP: Creating a kubernetes client
... skipping 2 lines ...
STEP: Waiting for a default service account to be provisioned in namespace
[BeforeEach] [sig-storage] Projected downwardAPI
  /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/common/storage/projected_downwardapi.go:41
[It] should set DefaultMode on files [LinuxOnly] [NodeConformance] [Conformance]
  /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:630
STEP: Creating a pod to test downward API volume plugin
Jul 18 19:52:39.040: INFO: Waiting up to 5m0s for pod "downwardapi-volume-d44b7a88-4a8d-4401-b349-6fb8627152a2" in namespace "projected-5668" to be "Succeeded or Failed"
Jul 18 19:52:39.075: INFO: Pod "downwardapi-volume-d44b7a88-4a8d-4401-b349-6fb8627152a2": Phase="Pending", Reason="", readiness=false. Elapsed: 35.080255ms
Jul 18 19:52:41.111: INFO: Pod "downwardapi-volume-d44b7a88-4a8d-4401-b349-6fb8627152a2": Phase="Pending", Reason="", readiness=false. Elapsed: 2.07063818s
Jul 18 19:52:43.144: INFO: Pod "downwardapi-volume-d44b7a88-4a8d-4401-b349-6fb8627152a2": Phase="Pending", Reason="", readiness=false. Elapsed: 4.103682164s
Jul 18 19:52:45.175: INFO: Pod "downwardapi-volume-d44b7a88-4a8d-4401-b349-6fb8627152a2": Phase="Succeeded", Reason="", readiness=false. Elapsed: 6.135187144s
STEP: Saw pod success
Jul 18 19:52:45.175: INFO: Pod "downwardapi-volume-d44b7a88-4a8d-4401-b349-6fb8627152a2" satisfied condition "Succeeded or Failed"
Jul 18 19:52:45.206: INFO: Trying to get logs from node 10.116.0.5 pod downwardapi-volume-d44b7a88-4a8d-4401-b349-6fb8627152a2 container client-container: <nil>
STEP: delete the pod
Jul 18 19:52:45.294: INFO: Waiting for pod downwardapi-volume-d44b7a88-4a8d-4401-b349-6fb8627152a2 to disappear
Jul 18 19:52:45.324: INFO: Pod downwardapi-volume-d44b7a88-4a8d-4401-b349-6fb8627152a2 no longer exists
[AfterEach] [sig-storage] Projected downwardAPI
  /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:186
... skipping 4 lines ...
• [SLOW TEST:6.573 seconds]
[sig-storage] Projected downwardAPI
/workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/common/storage/framework.go:23
  should set DefaultMode on files [LinuxOnly] [NodeConformance] [Conformance]
  /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:630
------------------------------
{"msg":"PASSED [sig-storage] Projected downwardAPI should set DefaultMode on files [LinuxOnly] [NodeConformance] [Conformance]","total":-1,"completed":4,"skipped":46,"failed":0}
[BeforeEach] [Testpattern: Inline-volume (ext3)] volumes
  /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/storage/framework/testsuite.go:51
Jul 18 19:52:45.430: INFO: Only supported for providers [azure] (not skeleton)
[AfterEach] [Testpattern: Inline-volume (ext3)] volumes
  /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:186

... skipping 103 lines ...
• [SLOW TEST:13.636 seconds]
[sig-api-machinery] ResourceQuota
/workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/apimachinery/framework.go:23
  should create a ResourceQuota and capture the life of a pod. [Conformance]
  /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:630
------------------------------
{"msg":"PASSED [sig-api-machinery] ResourceQuota should create a ResourceQuota and capture the life of a pod. [Conformance]","total":-1,"completed":5,"skipped":21,"failed":0}
[BeforeEach] [Testpattern: Dynamic PV (default fs)] subPath
  /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/storage/framework/testsuite.go:51
Jul 18 19:52:46.281: INFO: Only supported for providers [vsphere] (not skeleton)
[AfterEach] [Testpattern: Dynamic PV (default fs)] subPath
  /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:186

... skipping 171 lines ...
Jul 18 19:52:36.259: INFO: PersistentVolumeClaim pvc-dhvb4 found but phase is Pending instead of Bound.
Jul 18 19:52:38.291: INFO: PersistentVolumeClaim pvc-dhvb4 found and phase=Bound (14.25363061s)
Jul 18 19:52:38.291: INFO: Waiting up to 3m0s for PersistentVolume local-jh446 to have phase Bound
Jul 18 19:52:38.324: INFO: PersistentVolume local-jh446 found and phase=Bound (33.145992ms)
STEP: Creating pod pod-subpath-test-preprovisionedpv-l2xh
STEP: Creating a pod to test subpath
Jul 18 19:52:38.428: INFO: Waiting up to 5m0s for pod "pod-subpath-test-preprovisionedpv-l2xh" in namespace "provisioning-2688" to be "Succeeded or Failed"
Jul 18 19:52:38.461: INFO: Pod "pod-subpath-test-preprovisionedpv-l2xh": Phase="Pending", Reason="", readiness=false. Elapsed: 32.432514ms
Jul 18 19:52:40.495: INFO: Pod "pod-subpath-test-preprovisionedpv-l2xh": Phase="Pending", Reason="", readiness=false. Elapsed: 2.06642113s
Jul 18 19:52:42.528: INFO: Pod "pod-subpath-test-preprovisionedpv-l2xh": Phase="Pending", Reason="", readiness=false. Elapsed: 4.100073515s
Jul 18 19:52:44.560: INFO: Pod "pod-subpath-test-preprovisionedpv-l2xh": Phase="Pending", Reason="", readiness=false. Elapsed: 6.131585502s
Jul 18 19:52:46.592: INFO: Pod "pod-subpath-test-preprovisionedpv-l2xh": Phase="Pending", Reason="", readiness=false. Elapsed: 8.163415688s
Jul 18 19:52:48.625: INFO: Pod "pod-subpath-test-preprovisionedpv-l2xh": Phase="Pending", Reason="", readiness=false. Elapsed: 10.197144222s
Jul 18 19:52:50.658: INFO: Pod "pod-subpath-test-preprovisionedpv-l2xh": Phase="Pending", Reason="", readiness=false. Elapsed: 12.230013988s
Jul 18 19:52:52.692: INFO: Pod "pod-subpath-test-preprovisionedpv-l2xh": Phase="Succeeded", Reason="", readiness=false. Elapsed: 14.263717832s
STEP: Saw pod success
Jul 18 19:52:52.692: INFO: Pod "pod-subpath-test-preprovisionedpv-l2xh" satisfied condition "Succeeded or Failed"
Jul 18 19:52:52.724: INFO: Trying to get logs from node 10.116.0.5 pod pod-subpath-test-preprovisionedpv-l2xh container test-container-volume-preprovisionedpv-l2xh: <nil>
STEP: delete the pod
Jul 18 19:52:52.815: INFO: Waiting for pod pod-subpath-test-preprovisionedpv-l2xh to disappear
Jul 18 19:52:52.846: INFO: Pod pod-subpath-test-preprovisionedpv-l2xh no longer exists
STEP: Deleting pod pod-subpath-test-preprovisionedpv-l2xh
Jul 18 19:52:52.846: INFO: Deleting pod "pod-subpath-test-preprovisionedpv-l2xh" in namespace "provisioning-2688"
... skipping 19 lines ...
  /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/storage/in_tree_volumes.go:58
    [Testpattern: Pre-provisioned PV (default fs)] subPath
    /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/storage/framework/testsuite.go:50
      should support existing directory
      /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/storage/testsuites/subpath.go:205
------------------------------
{"msg":"PASSED [sig-storage] In-tree Volumes [Driver: local][LocalVolumeType: dir-link-bindmounted] [Testpattern: Pre-provisioned PV (default fs)] subPath should support existing directory","total":-1,"completed":7,"skipped":48,"failed":0}

SS
------------------------------
[BeforeEach] [Testpattern: Pre-provisioned PV (ext4)] volumes
  /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/storage/framework/testsuite.go:51
Jul 18 19:52:53.379: INFO: Driver hostPathSymlink doesn't support PreprovisionedPV -- skipping
... skipping 109 lines ...
/workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/common/node/framework.go:23
  when create a pod with lifecycle hook
  /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/common/node/lifecycle_hook.go:43
    should execute prestop exec hook properly [NodeConformance] [Conformance]
    /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:630
------------------------------
{"msg":"PASSED [sig-node] Container Lifecycle Hook when create a pod with lifecycle hook should execute prestop exec hook properly [NodeConformance] [Conformance]","total":-1,"completed":8,"skipped":56,"failed":0}

SSSSSSSSSSSSSSSSS
------------------------------
[BeforeEach] [sig-node] Container Runtime
  /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:185
STEP: Creating a kubernetes client
... skipping 18 lines ...
  /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/common/node/runtime.go:41
    when running a container with a new image
    /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/common/node/runtime.go:266
      should not be able to pull image from invalid registry [NodeConformance]
      /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/common/node/runtime.go:377
------------------------------
{"msg":"PASSED [sig-network] Netpol API should support creating NetworkPolicy API operations","total":-1,"completed":3,"skipped":34,"failed":0}
[BeforeEach] [sig-node] Security Context
  /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:185
STEP: Creating a kubernetes client
Jul 18 19:52:39.636: INFO: >>> kubeConfig: /root/.kube/config
STEP: Building a namespace api object, basename security-context-test
STEP: Waiting for a default service account to be provisioned in namespace
[BeforeEach] [sig-node] Security Context
  /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/common/node/security_context.go:46
[It] should run the container as unprivileged when false [LinuxOnly] [NodeConformance] [Conformance]
  /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:630
Jul 18 19:52:40.017: INFO: Waiting up to 5m0s for pod "busybox-privileged-false-c4ccf091-0011-463f-9a4f-da036acbf696" in namespace "security-context-test-5204" to be "Succeeded or Failed"
Jul 18 19:52:40.049: INFO: Pod "busybox-privileged-false-c4ccf091-0011-463f-9a4f-da036acbf696": Phase="Pending", Reason="", readiness=false. Elapsed: 31.054868ms
Jul 18 19:52:42.081: INFO: Pod "busybox-privileged-false-c4ccf091-0011-463f-9a4f-da036acbf696": Phase="Pending", Reason="", readiness=false. Elapsed: 2.063418343s
Jul 18 19:52:44.114: INFO: Pod "busybox-privileged-false-c4ccf091-0011-463f-9a4f-da036acbf696": Phase="Pending", Reason="", readiness=false. Elapsed: 4.09605529s
Jul 18 19:52:46.147: INFO: Pod "busybox-privileged-false-c4ccf091-0011-463f-9a4f-da036acbf696": Phase="Pending", Reason="", readiness=false. Elapsed: 6.129677803s
Jul 18 19:52:48.179: INFO: Pod "busybox-privileged-false-c4ccf091-0011-463f-9a4f-da036acbf696": Phase="Pending", Reason="", readiness=false. Elapsed: 8.162018769s
Jul 18 19:52:50.213: INFO: Pod "busybox-privileged-false-c4ccf091-0011-463f-9a4f-da036acbf696": Phase="Pending", Reason="", readiness=false. Elapsed: 10.195832958s
Jul 18 19:52:52.248: INFO: Pod "busybox-privileged-false-c4ccf091-0011-463f-9a4f-da036acbf696": Phase="Pending", Reason="", readiness=false. Elapsed: 12.230558768s
Jul 18 19:52:54.281: INFO: Pod "busybox-privileged-false-c4ccf091-0011-463f-9a4f-da036acbf696": Phase="Succeeded", Reason="", readiness=false. Elapsed: 14.263362276s
Jul 18 19:52:54.281: INFO: Pod "busybox-privileged-false-c4ccf091-0011-463f-9a4f-da036acbf696" satisfied condition "Succeeded or Failed"
Jul 18 19:52:54.317: INFO: Got logs for pod "busybox-privileged-false-c4ccf091-0011-463f-9a4f-da036acbf696": "ip: RTNETLINK answers: Operation not permitted\n"
[AfterEach] [sig-node] Security Context
  /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:186
Jul 18 19:52:54.317: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready
STEP: Destroying namespace "security-context-test-5204" for this suite.

... skipping 3 lines ...
/workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/common/node/framework.go:23
  When creating a pod with privileged
  /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/common/node/security_context.go:232
    should run the container as unprivileged when false [LinuxOnly] [NodeConformance] [Conformance]
    /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:630
------------------------------
{"msg":"PASSED [sig-node] Security Context When creating a pod with privileged should run the container as unprivileged when false [LinuxOnly] [NodeConformance] [Conformance]","total":-1,"completed":4,"skipped":34,"failed":0}

S
------------------------------
{"msg":"PASSED [sig-node] Container Runtime blackbox test when running a container with a new image should not be able to pull image from invalid registry [NodeConformance]","total":-1,"completed":5,"skipped":28,"failed":0}

S
------------------------------
[BeforeEach] [Testpattern: Inline-volume (default fs)] subPath
  /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/storage/framework/testsuite.go:51
Jul 18 19:52:54.427: INFO: Only supported for providers [aws] (not skeleton)
... skipping 249 lines ...
Jul 18 19:52:40.183: INFO: >>> kubeConfig: /root/.kube/config
STEP: Building a namespace api object, basename security-context
STEP: Waiting for a default service account to be provisioned in namespace
[It] should support seccomp unconfined on the container [LinuxOnly]
  /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/node/security_context.go:161
STEP: Creating a pod to test seccomp.security.alpha.kubernetes.io/pod
Jul 18 19:52:40.372: INFO: Waiting up to 5m0s for pod "security-context-10039c6f-0ec8-41b5-b79f-4656269ee8da" in namespace "security-context-2897" to be "Succeeded or Failed"
Jul 18 19:52:40.403: INFO: Pod "security-context-10039c6f-0ec8-41b5-b79f-4656269ee8da": Phase="Pending", Reason="", readiness=false. Elapsed: 30.649239ms
Jul 18 19:52:42.434: INFO: Pod "security-context-10039c6f-0ec8-41b5-b79f-4656269ee8da": Phase="Pending", Reason="", readiness=false. Elapsed: 2.062211031s
Jul 18 19:52:44.466: INFO: Pod "security-context-10039c6f-0ec8-41b5-b79f-4656269ee8da": Phase="Pending", Reason="", readiness=false. Elapsed: 4.093815823s
Jul 18 19:52:46.498: INFO: Pod "security-context-10039c6f-0ec8-41b5-b79f-4656269ee8da": Phase="Pending", Reason="", readiness=false. Elapsed: 6.126218864s
Jul 18 19:52:48.531: INFO: Pod "security-context-10039c6f-0ec8-41b5-b79f-4656269ee8da": Phase="Pending", Reason="", readiness=false. Elapsed: 8.158759671s
Jul 18 19:52:50.563: INFO: Pod "security-context-10039c6f-0ec8-41b5-b79f-4656269ee8da": Phase="Pending", Reason="", readiness=false. Elapsed: 10.190741348s
Jul 18 19:52:52.596: INFO: Pod "security-context-10039c6f-0ec8-41b5-b79f-4656269ee8da": Phase="Pending", Reason="", readiness=false. Elapsed: 12.223587359s
Jul 18 19:52:54.628: INFO: Pod "security-context-10039c6f-0ec8-41b5-b79f-4656269ee8da": Phase="Succeeded", Reason="", readiness=false. Elapsed: 14.255615362s
STEP: Saw pod success
Jul 18 19:52:54.628: INFO: Pod "security-context-10039c6f-0ec8-41b5-b79f-4656269ee8da" satisfied condition "Succeeded or Failed"
Jul 18 19:52:54.659: INFO: Trying to get logs from node 10.116.0.7 pod security-context-10039c6f-0ec8-41b5-b79f-4656269ee8da container test-container: <nil>
STEP: delete the pod
Jul 18 19:52:54.732: INFO: Waiting for pod security-context-10039c6f-0ec8-41b5-b79f-4656269ee8da to disappear
Jul 18 19:52:54.763: INFO: Pod security-context-10039c6f-0ec8-41b5-b79f-4656269ee8da no longer exists
[AfterEach] [sig-node] Security Context
  /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:186
... skipping 4 lines ...
• [SLOW TEST:14.673 seconds]
[sig-node] Security Context
/workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/node/framework.go:23
  should support seccomp unconfined on the container [LinuxOnly]
  /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/node/security_context.go:161
------------------------------
{"msg":"PASSED [sig-node] Security Context should support seccomp unconfined on the container [LinuxOnly]","total":-1,"completed":4,"skipped":29,"failed":0}

SS
------------------------------
[BeforeEach] [sig-storage] PersistentVolumes-local 
  /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:185
STEP: Creating a kubernetes client
... skipping 65 lines ...
  /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/storage/persistent_volumes-local.go:192
    One pod requesting one prebound PVC
    /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/storage/persistent_volumes-local.go:209
      should be able to mount volume and write from pod1
      /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/storage/persistent_volumes-local.go:238
------------------------------
{"msg":"PASSED [sig-storage] PersistentVolumes-local  [Volume type: block] One pod requesting one prebound PVC should be able to mount volume and write from pod1","total":-1,"completed":6,"skipped":87,"failed":0}

S
------------------------------
[BeforeEach] [Testpattern: Dynamic PV (block volmode)] volumes
  /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/storage/framework/testsuite.go:51
Jul 18 19:52:56.977: INFO: Only supported for providers [gce gke] (not skeleton)
... skipping 60 lines ...
  /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:186
Jul 18 19:52:57.410: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready
STEP: Destroying namespace "replication-controller-8672" for this suite.

•
------------------------------
{"msg":"PASSED [sig-apps] ReplicationController should surface a failure condition on a common issue like exceeded quota [Conformance]","total":-1,"completed":7,"skipped":97,"failed":0}

SS
------------------------------
[BeforeEach] [Testpattern: Dynamic PV (default fs)] subPath
  /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/storage/framework/testsuite.go:51
Jul 18 19:52:57.495: INFO: Driver hostPathSymlink doesn't support DynamicPV -- skipping
... skipping 117 lines ...
  /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:186
Jul 18 19:52:57.778: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready
STEP: Destroying namespace "tables-9744" for this suite.

•
------------------------------
{"msg":"PASSED [sig-api-machinery] Servers with support for Table transformation should return a 406 for a backend which does not implement metadata [Conformance]","total":-1,"completed":8,"skipped":111,"failed":0}

SSS
------------------------------
[BeforeEach] [sig-apps] DisruptionController
  /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:185
STEP: Creating a kubernetes client
... skipping 20 lines ...
• [SLOW TEST:18.603 seconds]
[sig-apps] DisruptionController
/workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/apps/framework.go:23
  evictions: enough pods, absolute => should allow an eviction
  /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/apps/disruption.go:267
------------------------------
{"msg":"PASSED [sig-apps] DisruptionController evictions: enough pods, absolute =\u003e should allow an eviction","total":-1,"completed":5,"skipped":56,"failed":0}

S
------------------------------
[BeforeEach] [Testpattern: Inline-volume (ext3)] volumes
  /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/storage/framework/testsuite.go:51
Jul 18 19:53:04.091: INFO: Only supported for node OS distro [gci ubuntu custom] (not debian)
... skipping 162 lines ...
• [SLOW TEST:97.952 seconds]
[sig-network] Services
/workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/network/common/framework.go:23
  should be able to switch session affinity for service with type clusterIP [LinuxOnly] [Conformance]
  /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:630
------------------------------
{"msg":"PASSED [sig-network] Services should be able to switch session affinity for service with type clusterIP [LinuxOnly] [Conformance]","total":-1,"completed":2,"skipped":11,"failed":0}

S
------------------------------
[BeforeEach] [sig-network] Services
  /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:185
STEP: Creating a kubernetes client
... skipping 60 lines ...
• [SLOW TEST:47.479 seconds]
[sig-network] Services
/workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/network/common/framework.go:23
  should have session affinity work for NodePort service [LinuxOnly] [Conformance]
  /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:630
------------------------------
{"msg":"PASSED [sig-network] Services should have session affinity work for NodePort service [LinuxOnly] [Conformance]","total":-1,"completed":4,"skipped":23,"failed":0}
[BeforeEach] [Testpattern: Dynamic PV (block volmode)] provisioning
  /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/storage/framework/testsuite.go:51
Jul 18 19:53:05.503: INFO: Driver local doesn't support DynamicPV -- skipping
[AfterEach] [Testpattern: Dynamic PV (block volmode)] provisioning
  /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:186

... skipping 25 lines ...
[It] should support readOnly directory specified in the volumeMount
  /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/storage/testsuites/subpath.go:364
Jul 18 19:52:41.263: INFO: In-tree plugin kubernetes.io/host-path is not migrated, not validating any metrics
Jul 18 19:52:41.296: INFO: Creating resource for inline volume
STEP: Creating pod pod-subpath-test-inlinevolume-nkk8
STEP: Creating a pod to test subpath
Jul 18 19:52:41.338: INFO: Waiting up to 5m0s for pod "pod-subpath-test-inlinevolume-nkk8" in namespace "provisioning-1389" to be "Succeeded or Failed"
Jul 18 19:52:41.370: INFO: Pod "pod-subpath-test-inlinevolume-nkk8": Phase="Pending", Reason="", readiness=false. Elapsed: 31.601741ms
Jul 18 19:52:43.403: INFO: Pod "pod-subpath-test-inlinevolume-nkk8": Phase="Pending", Reason="", readiness=false. Elapsed: 2.064815437s
Jul 18 19:52:45.434: INFO: Pod "pod-subpath-test-inlinevolume-nkk8": Phase="Pending", Reason="", readiness=false. Elapsed: 4.095970934s
Jul 18 19:52:47.468: INFO: Pod "pod-subpath-test-inlinevolume-nkk8": Phase="Pending", Reason="", readiness=false. Elapsed: 6.129619163s
Jul 18 19:52:49.500: INFO: Pod "pod-subpath-test-inlinevolume-nkk8": Phase="Pending", Reason="", readiness=false. Elapsed: 8.161828713s
Jul 18 19:52:51.534: INFO: Pod "pod-subpath-test-inlinevolume-nkk8": Phase="Pending", Reason="", readiness=false. Elapsed: 10.195500522s
... skipping 2 lines ...
Jul 18 19:52:57.629: INFO: Pod "pod-subpath-test-inlinevolume-nkk8": Phase="Pending", Reason="", readiness=false. Elapsed: 16.290350698s
Jul 18 19:52:59.661: INFO: Pod "pod-subpath-test-inlinevolume-nkk8": Phase="Pending", Reason="", readiness=false. Elapsed: 18.322110618s
Jul 18 19:53:01.694: INFO: Pod "pod-subpath-test-inlinevolume-nkk8": Phase="Pending", Reason="", readiness=false. Elapsed: 20.355177523s
Jul 18 19:53:03.726: INFO: Pod "pod-subpath-test-inlinevolume-nkk8": Phase="Pending", Reason="", readiness=false. Elapsed: 22.387591271s
Jul 18 19:53:05.758: INFO: Pod "pod-subpath-test-inlinevolume-nkk8": Phase="Succeeded", Reason="", readiness=false. Elapsed: 24.419956997s
STEP: Saw pod success
Jul 18 19:53:05.758: INFO: Pod "pod-subpath-test-inlinevolume-nkk8" satisfied condition "Succeeded or Failed"
Jul 18 19:53:05.789: INFO: Trying to get logs from node 10.116.0.5 pod pod-subpath-test-inlinevolume-nkk8 container test-container-subpath-inlinevolume-nkk8: <nil>
STEP: delete the pod
Jul 18 19:53:05.867: INFO: Waiting for pod pod-subpath-test-inlinevolume-nkk8 to disappear
Jul 18 19:53:05.904: INFO: Pod pod-subpath-test-inlinevolume-nkk8 no longer exists
STEP: Deleting pod pod-subpath-test-inlinevolume-nkk8
Jul 18 19:53:05.904: INFO: Deleting pod "pod-subpath-test-inlinevolume-nkk8" in namespace "provisioning-1389"
... skipping 12 lines ...
  /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/storage/in_tree_volumes.go:58
    [Testpattern: Inline-volume (default fs)] subPath
    /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/storage/framework/testsuite.go:50
      should support readOnly directory specified in the volumeMount
      /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/storage/testsuites/subpath.go:364
------------------------------
{"msg":"PASSED [sig-storage] In-tree Volumes [Driver: hostPath] [Testpattern: Inline-volume (default fs)] subPath should support readOnly directory specified in the volumeMount","total":-1,"completed":7,"skipped":38,"failed":0}

S
------------------------------
[BeforeEach] [sig-network] NetworkPolicy API
  /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:185
STEP: Creating a kubernetes client
... skipping 24 lines ...
  /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:186
Jul 18 19:53:06.307: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready
STEP: Destroying namespace "networkpolicies-194" for this suite.

•
------------------------------
{"msg":"PASSED [sig-network] NetworkPolicy API should support creating NetworkPolicy API operations","total":-1,"completed":5,"skipped":27,"failed":0}

SSS
------------------------------
[BeforeEach] [sig-apps] StatefulSet
  /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:185
STEP: Creating a kubernetes client
... skipping 5 lines ...
[BeforeEach] Basic StatefulSet functionality [StatefulSetBasic]
  /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/apps/statefulset.go:105
STEP: Creating service test in namespace statefulset-3141
[It] should adopt matching orphans and release non-matching pods
  /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/apps/statefulset.go:165
STEP: Creating statefulset ss in namespace statefulset-3141
Jul 18 19:53:06.295: INFO: error finding default storageClass : No default storage class found
[AfterEach] Basic StatefulSet functionality [StatefulSetBasic]
  /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/apps/statefulset.go:116
Jul 18 19:53:06.297: INFO: Deleting all statefulset in ns statefulset-3141
[AfterEach] [sig-apps] StatefulSet
  /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:186
Jul 18 19:53:06.393: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready
... skipping 5 lines ...
/workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/apps/framework.go:23
  Basic StatefulSet functionality [StatefulSetBasic]
  /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/apps/statefulset.go:95
    should adopt matching orphans and release non-matching pods [It]
    /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/apps/statefulset.go:165

    error finding default storageClass : No default storage class found

    /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/pv/pv.go:819
------------------------------
SSSSS
------------------------------
[BeforeEach] [Testpattern: Dynamic PV (block volmode)] provisioning
... skipping 16 lines ...
      Driver local doesn't support DynamicPV -- skipping

      /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/storage/framework/testsuite.go:116
------------------------------
S
------------------------------
{"msg":"PASSED [sig-api-machinery] AdmissionWebhook [Privileged:ClusterAdmin] should unconditionally reject operations on fail closed webhook [Conformance]","total":-1,"completed":5,"skipped":13,"failed":0}
[BeforeEach] [sig-cli] Kubectl Port forwarding
  /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:185
STEP: Creating a kubernetes client
Jul 18 19:52:34.575: INFO: >>> kubeConfig: /root/.kube/config
STEP: Building a namespace api object, basename port-forwarding
STEP: Waiting for a default service account to be provisioned in namespace
... skipping 42 lines ...
  /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/kubectl/portforward.go:474
    that expects a client request
    /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/kubectl/portforward.go:475
      should support a client that connects, sends NO DATA, and disconnects
      /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/kubectl/portforward.go:476
------------------------------
{"msg":"PASSED [sig-cli] Kubectl Port forwarding With a server listening on localhost that expects a client request should support a client that connects, sends NO DATA, and disconnects","total":-1,"completed":6,"skipped":13,"failed":0}

SS
------------------------------
[BeforeEach] [sig-node] Pods
  /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:185
STEP: Creating a kubernetes client
... skipping 25 lines ...
• [SLOW TEST:16.480 seconds]
[sig-node] Pods
/workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/common/node/framework.go:23
  should support retrieving logs from the container over websockets [NodeConformance] [Conformance]
  /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:630
------------------------------
{"msg":"PASSED [sig-node] Pods should support retrieving logs from the container over websockets [NodeConformance] [Conformance]","total":-1,"completed":5,"skipped":31,"failed":0}

S
------------------------------
[BeforeEach] [Testpattern: Pre-provisioned PV (block volmode)] volumeMode
  /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/storage/framework/testsuite.go:51
Jul 18 19:53:11.372: INFO: Driver hostPathSymlink doesn't support PreprovisionedPV -- skipping
... skipping 98 lines ...
  /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/storage/persistent_volumes-local.go:192
    One pod requesting one prebound PVC
    /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/storage/persistent_volumes-local.go:209
      should be able to mount volume and read from pod1
      /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/storage/persistent_volumes-local.go:232
------------------------------
{"msg":"PASSED [sig-storage] PersistentVolumes-local  [Volume type: dir-link] One pod requesting one prebound PVC should be able to mount volume and read from pod1","total":-1,"completed":4,"skipped":21,"failed":0}

SSS
------------------------------
[BeforeEach] [sig-storage] HostPath
  /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:185
STEP: Creating a kubernetes client
... skipping 2 lines ...
STEP: Waiting for a default service account to be provisioned in namespace
[BeforeEach] [sig-storage] HostPath
  /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/common/storage/host_path.go:37
[It] should support subPath [NodeConformance]
  /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/common/storage/host_path.go:93
STEP: Creating a pod to test hostPath subPath
Jul 18 19:52:54.962: INFO: Waiting up to 5m0s for pod "pod-host-path-test" in namespace "hostpath-7507" to be "Succeeded or Failed"
Jul 18 19:52:54.994: INFO: Pod "pod-host-path-test": Phase="Pending", Reason="", readiness=false. Elapsed: 31.334696ms
Jul 18 19:52:57.026: INFO: Pod "pod-host-path-test": Phase="Pending", Reason="", readiness=false. Elapsed: 2.063682957s
Jul 18 19:52:59.062: INFO: Pod "pod-host-path-test": Phase="Pending", Reason="", readiness=false. Elapsed: 4.099312057s
Jul 18 19:53:01.095: INFO: Pod "pod-host-path-test": Phase="Pending", Reason="", readiness=false. Elapsed: 6.132347427s
Jul 18 19:53:03.127: INFO: Pod "pod-host-path-test": Phase="Pending", Reason="", readiness=false. Elapsed: 8.165234199s
Jul 18 19:53:05.162: INFO: Pod "pod-host-path-test": Phase="Pending", Reason="", readiness=false. Elapsed: 10.199720272s
Jul 18 19:53:07.195: INFO: Pod "pod-host-path-test": Phase="Pending", Reason="", readiness=false. Elapsed: 12.232815218s
Jul 18 19:53:09.230: INFO: Pod "pod-host-path-test": Phase="Pending", Reason="", readiness=false. Elapsed: 14.267589085s
Jul 18 19:53:11.261: INFO: Pod "pod-host-path-test": Phase="Succeeded", Reason="", readiness=false. Elapsed: 16.298968655s
STEP: Saw pod success
Jul 18 19:53:11.261: INFO: Pod "pod-host-path-test" satisfied condition "Succeeded or Failed"
Jul 18 19:53:11.293: INFO: Trying to get logs from node 10.116.0.7 pod pod-host-path-test container test-container-2: <nil>
STEP: delete the pod
Jul 18 19:53:11.373: INFO: Waiting for pod pod-host-path-test to disappear
Jul 18 19:53:11.405: INFO: Pod pod-host-path-test no longer exists
[AfterEach] [sig-storage] HostPath
  /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:186
... skipping 4 lines ...
• [SLOW TEST:16.741 seconds]
[sig-storage] HostPath
/workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/common/storage/framework.go:23
  should support subPath [NodeConformance]
  /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/common/storage/host_path.go:93
------------------------------
{"msg":"PASSED [sig-storage] HostPath should support subPath [NodeConformance]","total":-1,"completed":5,"skipped":54,"failed":0}

S
------------------------------
[BeforeEach] [Testpattern: Pre-provisioned PV (default fs)] subPath
  /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/storage/framework/testsuite.go:51
Jul 18 19:53:11.517: INFO: Only supported for providers [vsphere] (not skeleton)
... skipping 75 lines ...
  /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:186
Jul 18 19:53:12.182: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready
STEP: Destroying namespace "podtemplate-1149" for this suite.

•
------------------------------
{"msg":"PASSED [sig-node] PodTemplates should run the lifecycle of PodTemplates [Conformance]","total":-1,"completed":6,"skipped":68,"failed":0}
[BeforeEach] [Testpattern: Dynamic PV (default fs)] fsgroupchangepolicy
  /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/storage/framework/testsuite.go:51
Jul 18 19:53:12.259: INFO: Only supported for providers [vsphere] (not skeleton)
[AfterEach] [Testpattern: Dynamic PV (default fs)] fsgroupchangepolicy
  /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:186

... skipping 153 lines ...
/workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/storage/utils/framework.go:23
  CSIStorageCapacity
  /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/storage/csi_mock_volume.go:1134
    CSIStorageCapacity used, insufficient capacity
    /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/storage/csi_mock_volume.go:1177
------------------------------
{"msg":"PASSED [sig-storage] CSI mock volume CSIStorageCapacity CSIStorageCapacity used, insufficient capacity","total":-1,"completed":5,"skipped":18,"failed":0}

S
------------------------------
[BeforeEach] [sig-storage] CSI mock volume
  /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:185
STEP: Creating a kubernetes client
... skipping 98 lines ...
/workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/storage/utils/framework.go:23
  CSIStorageCapacity
  /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/storage/csi_mock_volume.go:1134
    CSIStorageCapacity used, no capacity
    /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/storage/csi_mock_volume.go:1177
------------------------------
{"msg":"PASSED [sig-storage] CSI mock volume CSIStorageCapacity CSIStorageCapacity used, no capacity","total":-1,"completed":3,"skipped":10,"failed":0}

SS
------------------------------
[BeforeEach] [Testpattern: Pre-provisioned PV (ext4)] volumes
  /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/storage/framework/testsuite.go:51
Jul 18 19:53:13.981: INFO: Driver local doesn't support ext4 -- skipping
... skipping 160 lines ...
/workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/apimachinery/framework.go:23
  Simple CustomResourceDefinition
  /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/apimachinery/custom_resource_definition.go:48
    listing custom resource definition objects works  [Conformance]
    /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:630
------------------------------
{"msg":"PASSED [sig-api-machinery] CustomResourceDefinition resources [Privileged:ClusterAdmin] Simple CustomResourceDefinition listing custom resource definition objects works  [Conformance]","total":-1,"completed":3,"skipped":8,"failed":0}

SS
------------------------------
[BeforeEach] [Testpattern: Dynamic PV (default fs)] fsgroupchangepolicy
  /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/storage/framework/testsuite.go:51
Jul 18 19:53:14.617: INFO: Only supported for providers [gce gke] (not skeleton)
... skipping 165 lines ...
• [SLOW TEST:69.403 seconds]
[sig-network] Conntrack
/workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/network/common/framework.go:23
  should be able to preserve UDP traffic when server pod cycles for a NodePort service
  /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/network/conntrack.go:130
------------------------------
{"msg":"PASSED [sig-network] Conntrack should be able to preserve UDP traffic when server pod cycles for a NodePort service","total":-1,"completed":3,"skipped":21,"failed":0}
[BeforeEach] [Testpattern: Dynamic PV (default fs)] subPath
  /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/storage/framework/testsuite.go:51
Jul 18 19:53:16.761: INFO: Only supported for node OS distro [gci ubuntu custom] (not debian)
[AfterEach] [Testpattern: Dynamic PV (default fs)] subPath
  /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:186

... skipping 34 lines ...
      Driver local doesn't support DynamicPV -- skipping

      /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/storage/framework/testsuite.go:116
------------------------------
S
------------------------------
{"msg":"PASSED [sig-instrumentation] MetricsGrabber should grab all metrics from API server.","total":-1,"completed":1,"skipped":5,"failed":0}
[BeforeEach] [sig-node] Probing container
  /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:185
STEP: Creating a kubernetes client
Jul 18 19:50:37.275: INFO: >>> kubeConfig: /root/.kube/config
STEP: Building a namespace api object, basename container-probe
STEP: Waiting for a default service account to be provisioned in namespace
... skipping 20 lines ...
• [SLOW TEST:159.890 seconds]
[sig-node] Probing container
/workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/common/node/framework.go:23
  should have monotonically increasing restart count [NodeConformance] [Conformance]
  /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:630
------------------------------
{"msg":"PASSED [sig-node] Probing container should have monotonically increasing restart count [NodeConformance] [Conformance]","total":-1,"completed":2,"skipped":5,"failed":0}

SSS
------------------------------
[BeforeEach] [Testpattern: Pre-provisioned PV (block volmode)] volumeMode
  /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/storage/framework/testsuite.go:51
[BeforeEach] [Testpattern: Pre-provisioned PV (block volmode)] volumeMode
... skipping 37 lines ...
[AfterEach] [sig-api-machinery] client-go should negotiate
  /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:186
Jul 18 19:53:17.484: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready

•
------------------------------
{"msg":"PASSED [sig-api-machinery] client-go should negotiate watch and report errors with accept \"application/vnd.kubernetes.protobuf,application/json\"","total":-1,"completed":3,"skipped":10,"failed":0}

SSS
------------------------------
[BeforeEach] [sig-cli] Kubectl client
  /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:185
STEP: Creating a kubernetes client
... skipping 25 lines ...
  /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:186
Jul 18 19:53:18.967: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready
STEP: Destroying namespace "kubectl-1283" for this suite.

•
------------------------------
{"msg":"PASSED [sig-cli] Kubectl client Kubectl apply should reuse port when apply to an existing SVC","total":-1,"completed":4,"skipped":13,"failed":0}

SSSSS
------------------------------
[BeforeEach] [sig-node] Probing container
  /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:185
STEP: Creating a kubernetes client
... skipping 19 lines ...
• [SLOW TEST:67.825 seconds]
[sig-node] Probing container
/workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/common/node/framework.go:23
  should be restarted by liveness probe after startup probe enables it
  /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/common/node/container_probe.go:371
------------------------------
{"msg":"PASSED [sig-node] Probing container should be restarted by liveness probe after startup probe enables it","total":-1,"completed":4,"skipped":25,"failed":0}

S
------------------------------
[BeforeEach] [Testpattern: Inline-volume (ext4)] volumes
  /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/storage/framework/testsuite.go:51
Jul 18 19:53:20.671: INFO: Only supported for providers [openstack] (not skeleton)
... skipping 46 lines ...
STEP: Waiting for a default service account to be provisioned in namespace
[BeforeEach] [sig-storage] Projected downwardAPI
  /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/common/storage/projected_downwardapi.go:41
[It] should provide node allocatable (cpu) as default cpu limit if the limit is not set [NodeConformance] [Conformance]
  /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:630
STEP: Creating a pod to test downward API volume plugin
Jul 18 19:53:11.979: INFO: Waiting up to 5m0s for pod "downwardapi-volume-56b66443-b600-4525-8a36-e24f1d654f9d" in namespace "projected-2453" to be "Succeeded or Failed"
Jul 18 19:53:12.011: INFO: Pod "downwardapi-volume-56b66443-b600-4525-8a36-e24f1d654f9d": Phase="Pending", Reason="", readiness=false. Elapsed: 32.702917ms
Jul 18 19:53:14.043: INFO: Pod "downwardapi-volume-56b66443-b600-4525-8a36-e24f1d654f9d": Phase="Pending", Reason="", readiness=false. Elapsed: 2.064465638s
Jul 18 19:53:16.081: INFO: Pod "downwardapi-volume-56b66443-b600-4525-8a36-e24f1d654f9d": Phase="Pending", Reason="", readiness=false. Elapsed: 4.101945749s
Jul 18 19:53:18.113: INFO: Pod "downwardapi-volume-56b66443-b600-4525-8a36-e24f1d654f9d": Phase="Pending", Reason="", readiness=false. Elapsed: 6.134059547s
Jul 18 19:53:20.144: INFO: Pod "downwardapi-volume-56b66443-b600-4525-8a36-e24f1d654f9d": Phase="Pending", Reason="", readiness=false. Elapsed: 8.1654274s
Jul 18 19:53:22.175: INFO: Pod "downwardapi-volume-56b66443-b600-4525-8a36-e24f1d654f9d": Phase="Succeeded", Reason="", readiness=false. Elapsed: 10.196659291s
STEP: Saw pod success
Jul 18 19:53:22.175: INFO: Pod "downwardapi-volume-56b66443-b600-4525-8a36-e24f1d654f9d" satisfied condition "Succeeded or Failed"
Jul 18 19:53:22.205: INFO: Trying to get logs from node 10.116.0.5 pod downwardapi-volume-56b66443-b600-4525-8a36-e24f1d654f9d container client-container: <nil>
STEP: delete the pod
Jul 18 19:53:22.276: INFO: Waiting for pod downwardapi-volume-56b66443-b600-4525-8a36-e24f1d654f9d to disappear
Jul 18 19:53:22.307: INFO: Pod downwardapi-volume-56b66443-b600-4525-8a36-e24f1d654f9d no longer exists
[AfterEach] [sig-storage] Projected downwardAPI
  /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:186
... skipping 4 lines ...
• [SLOW TEST:10.912 seconds]
[sig-storage] Projected downwardAPI
/workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/common/storage/framework.go:23
  should provide node allocatable (cpu) as default cpu limit if the limit is not set [NodeConformance] [Conformance]
  /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:630
------------------------------
{"msg":"PASSED [sig-storage] Projected downwardAPI should provide node allocatable (cpu) as default cpu limit if the limit is not set [NodeConformance] [Conformance]","total":-1,"completed":5,"skipped":24,"failed":0}

S
------------------------------
[BeforeEach] [Testpattern: Inline-volume (default fs)] subPath
  /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/storage/framework/testsuite.go:51
[BeforeEach] [Testpattern: Inline-volume (default fs)] subPath
... skipping 5 lines ...
[It] should support readOnly file specified in the volumeMount [LinuxOnly]
  /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/storage/testsuites/subpath.go:379
Jul 18 19:53:16.940: INFO: In-tree plugin kubernetes.io/host-path is not migrated, not validating any metrics
Jul 18 19:53:16.972: INFO: Creating resource for inline volume
STEP: Creating pod pod-subpath-test-inlinevolume-x9ht
STEP: Creating a pod to test subpath
Jul 18 19:53:17.010: INFO: Waiting up to 5m0s for pod "pod-subpath-test-inlinevolume-x9ht" in namespace "provisioning-8560" to be "Succeeded or Failed"
Jul 18 19:53:17.043: INFO: Pod "pod-subpath-test-inlinevolume-x9ht": Phase="Pending", Reason="", readiness=false. Elapsed: 32.973946ms
Jul 18 19:53:19.076: INFO: Pod "pod-subpath-test-inlinevolume-x9ht": Phase="Pending", Reason="", readiness=false. Elapsed: 2.065449961s
Jul 18 19:53:21.108: INFO: Pod "pod-subpath-test-inlinevolume-x9ht": Phase="Pending", Reason="", readiness=false. Elapsed: 4.097812662s
Jul 18 19:53:23.140: INFO: Pod "pod-subpath-test-inlinevolume-x9ht": Phase="Pending", Reason="", readiness=false. Elapsed: 6.130041819s
Jul 18 19:53:25.173: INFO: Pod "pod-subpath-test-inlinevolume-x9ht": Phase="Succeeded", Reason="", readiness=false. Elapsed: 8.162503905s
STEP: Saw pod success
Jul 18 19:53:25.173: INFO: Pod "pod-subpath-test-inlinevolume-x9ht" satisfied condition "Succeeded or Failed"
Jul 18 19:53:25.203: INFO: Trying to get logs from node 10.116.0.7 pod pod-subpath-test-inlinevolume-x9ht container test-container-subpath-inlinevolume-x9ht: <nil>
STEP: delete the pod
Jul 18 19:53:25.288: INFO: Waiting for pod pod-subpath-test-inlinevolume-x9ht to disappear
Jul 18 19:53:25.319: INFO: Pod pod-subpath-test-inlinevolume-x9ht no longer exists
STEP: Deleting pod pod-subpath-test-inlinevolume-x9ht
Jul 18 19:53:25.319: INFO: Deleting pod "pod-subpath-test-inlinevolume-x9ht" in namespace "provisioning-8560"
... skipping 12 lines ...
  /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/storage/in_tree_volumes.go:58
    [Testpattern: Inline-volume (default fs)] subPath
    /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/storage/framework/testsuite.go:50
      should support readOnly file specified in the volumeMount [LinuxOnly]
      /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/storage/testsuites/subpath.go:379
------------------------------
{"msg":"PASSED [sig-storage] In-tree Volumes [Driver: hostPath] [Testpattern: Inline-volume (default fs)] subPath should support readOnly file specified in the volumeMount [LinuxOnly]","total":-1,"completed":4,"skipped":25,"failed":0}

SS
------------------------------
[BeforeEach] [Testpattern: Inline-volume (default fs)] subPath
  /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/storage/framework/testsuite.go:51
Jul 18 19:53:25.498: INFO: Only supported for providers [openstack] (not skeleton)
... skipping 62 lines ...
  /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:186
Jul 18 19:53:25.936: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready
STEP: Destroying namespace "watch-3255" for this suite.

•
------------------------------
{"msg":"PASSED [sig-api-machinery] Watchers should be able to restart watching from the last resource version observed by the previous watch [Conformance]","total":-1,"completed":5,"skipped":30,"failed":0}
[BeforeEach] [Testpattern: Dynamic PV (default fs)] subPath
  /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/storage/framework/testsuite.go:51
Jul 18 19:53:26.013: INFO: Only supported for providers [azure] (not skeleton)
[AfterEach] [Testpattern: Dynamic PV (default fs)] subPath
  /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:186

... skipping 84 lines ...
  /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/storage/persistent_volumes-local.go:192
    One pod requesting one prebound PVC
    /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/storage/persistent_volumes-local.go:209
      should be able to mount volume and read from pod1
      /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/storage/persistent_volumes-local.go:232
------------------------------
{"msg":"PASSED [sig-storage] PersistentVolumes-local  [Volume type: tmpfs] One pod requesting one prebound PVC should be able to mount volume and read from pod1","total":-1,"completed":8,"skipped":47,"failed":0}

S
------------------------------
[BeforeEach] [Testpattern: Dynamic PV (default fs)] fsgroupchangepolicy
  /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/storage/framework/testsuite.go:51
Jul 18 19:53:30.864: INFO: Only supported for providers [azure] (not skeleton)
... skipping 34 lines ...
  /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:186
Jul 18 19:53:31.100: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready
STEP: Destroying namespace "kubectl-4057" for this suite.

•
------------------------------
{"msg":"PASSED [sig-cli] Kubectl client Kubectl create quota should reject quota with invalid scopes","total":-1,"completed":9,"skipped":50,"failed":0}
[BeforeEach] [Testpattern: Pre-provisioned PV (default fs)] subPath
  /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/storage/framework/testsuite.go:51
Jul 18 19:53:31.175: INFO: Only supported for providers [openstack] (not skeleton)
[AfterEach] [Testpattern: Pre-provisioned PV (default fs)] subPath
  /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:186

... skipping 42 lines ...
STEP: Building a namespace api object, basename security-context-test
STEP: Waiting for a default service account to be provisioned in namespace
[BeforeEach] [sig-node] Security Context
  /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/common/node/security_context.go:46
[It] should run the container with readonly rootfs when readOnlyRootFilesystem=true [LinuxOnly] [NodeConformance]
  /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/common/node/security_context.go:217
Jul 18 19:53:14.834: INFO: Waiting up to 5m0s for pod "busybox-readonly-true-59182990-5965-49fb-84a6-ab00c12846de" in namespace "security-context-test-6949" to be "Succeeded or Failed"
Jul 18 19:53:14.867: INFO: Pod "busybox-readonly-true-59182990-5965-49fb-84a6-ab00c12846de": Phase="Pending", Reason="", readiness=false. Elapsed: 33.092053ms
Jul 18 19:53:16.899: INFO: Pod "busybox-readonly-true-59182990-5965-49fb-84a6-ab00c12846de": Phase="Pending", Reason="", readiness=false. Elapsed: 2.064665914s
Jul 18 19:53:18.932: INFO: Pod "busybox-readonly-true-59182990-5965-49fb-84a6-ab00c12846de": Phase="Pending", Reason="", readiness=false. Elapsed: 4.097533073s
Jul 18 19:53:20.964: INFO: Pod "busybox-readonly-true-59182990-5965-49fb-84a6-ab00c12846de": Phase="Pending", Reason="", readiness=false. Elapsed: 6.129597783s
Jul 18 19:53:22.995: INFO: Pod "busybox-readonly-true-59182990-5965-49fb-84a6-ab00c12846de": Phase="Pending", Reason="", readiness=false. Elapsed: 8.160761091s
Jul 18 19:53:25.027: INFO: Pod "busybox-readonly-true-59182990-5965-49fb-84a6-ab00c12846de": Phase="Pending", Reason="", readiness=false. Elapsed: 10.19225075s
Jul 18 19:53:27.059: INFO: Pod "busybox-readonly-true-59182990-5965-49fb-84a6-ab00c12846de": Phase="Pending", Reason="", readiness=false. Elapsed: 12.224509295s
Jul 18 19:53:29.091: INFO: Pod "busybox-readonly-true-59182990-5965-49fb-84a6-ab00c12846de": Phase="Pending", Reason="", readiness=false. Elapsed: 14.256310928s
Jul 18 19:53:31.122: INFO: Pod "busybox-readonly-true-59182990-5965-49fb-84a6-ab00c12846de": Phase="Failed", Reason="", readiness=false. Elapsed: 16.287580356s
Jul 18 19:53:31.122: INFO: Pod "busybox-readonly-true-59182990-5965-49fb-84a6-ab00c12846de" satisfied condition "Succeeded or Failed"
[AfterEach] [sig-node] Security Context
  /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:186
Jul 18 19:53:31.122: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready
STEP: Destroying namespace "security-context-test-6949" for this suite.


... skipping 2 lines ...
/workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/common/node/framework.go:23
  When creating a pod with readOnlyRootFilesystem
  /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/common/node/security_context.go:171
    should run the container with readonly rootfs when readOnlyRootFilesystem=true [LinuxOnly] [NodeConformance]
    /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/common/node/security_context.go:217
------------------------------
{"msg":"PASSED [sig-node] Security Context When creating a pod with readOnlyRootFilesystem should run the container with readonly rootfs when readOnlyRootFilesystem=true [LinuxOnly] [NodeConformance]","total":-1,"completed":4,"skipped":14,"failed":0}
[BeforeEach] [Testpattern: Dynamic PV (default fs)] volumes
  /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/storage/framework/testsuite.go:51
Jul 18 19:53:31.222: INFO: Only supported for providers [gce gke] (not skeleton)
[AfterEach] [Testpattern: Dynamic PV (default fs)] volumes
  /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:186

... skipping 202 lines ...
/workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/network/common/framework.go:23
  should support configurable pod DNS nameservers [Conformance]
  /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:630
------------------------------
S
------------------------------
{"msg":"PASSED [sig-network] DNS should support configurable pod DNS nameservers [Conformance]","total":-1,"completed":6,"skipped":25,"failed":0}

S
------------------------------
[BeforeEach] [Testpattern: Pre-provisioned PV (ext3)] volumes
  /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/storage/framework/testsuite.go:51
Jul 18 19:53:31.327: INFO: Driver local doesn't support ext3 -- skipping
... skipping 77 lines ...
  /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/storage/in_tree_volumes.go:58
    [Testpattern: Pre-provisioned PV (filesystem volmode)] volumeMode
    /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/storage/framework/testsuite.go:50
      should not mount / map unused volumes in a pod [LinuxOnly]
      /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/storage/testsuites/volumemode.go:351
------------------------------
{"msg":"PASSED [sig-storage] In-tree Volumes [Driver: local][LocalVolumeType: dir-link] [Testpattern: Pre-provisioned PV (filesystem volmode)] volumeMode should not mount / map unused volumes in a pod [LinuxOnly]","total":-1,"completed":8,"skipped":29,"failed":0}

SSS
------------------------------
[BeforeEach] [Testpattern: Dynamic PV (ext4)] volumes
  /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/storage/framework/testsuite.go:51
Jul 18 19:53:34.605: INFO: Driver hostPathSymlink doesn't support DynamicPV -- skipping
... skipping 126 lines ...
• [SLOW TEST:88.601 seconds]
[sig-network] Services
/workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/network/common/framework.go:23
  should have session affinity timeout work for NodePort service [LinuxOnly] [Conformance]
  /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:630
------------------------------
{"msg":"PASSED [sig-network] Services should have session affinity timeout work for NodePort service [LinuxOnly] [Conformance]","total":-1,"completed":6,"skipped":47,"failed":0}
[BeforeEach] [Testpattern: Pre-provisioned PV (default fs)] subPath
  /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/storage/framework/testsuite.go:51
Jul 18 19:53:36.388: INFO: Only supported for providers [aws] (not skeleton)
[AfterEach] [Testpattern: Pre-provisioned PV (default fs)] subPath
  /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:186

... skipping 134 lines ...
  /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/storage/in_tree_volumes.go:58
    [Testpattern: Inline-volume (default fs)] subPath
    /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/storage/framework/testsuite.go:50
      should be able to unmount after the subpath directory is deleted [LinuxOnly]
      /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/storage/testsuites/subpath.go:444
------------------------------
{"msg":"PASSED [sig-storage] In-tree Volumes [Driver: emptydir] [Testpattern: Inline-volume (default fs)] subPath should be able to unmount after the subpath directory is deleted [LinuxOnly]","total":-1,"completed":4,"skipped":39,"failed":0}

SSSS
------------------------------
[BeforeEach] [Testpattern: Dynamic PV (ext3)] volumes
  /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/storage/framework/testsuite.go:51
Jul 18 19:53:37.332: INFO: Only supported for providers [aws] (not skeleton)
... skipping 21 lines ...
Jul 18 19:53:31.199: INFO: >>> kubeConfig: /root/.kube/config
STEP: Building a namespace api object, basename emptydir
STEP: Waiting for a default service account to be provisioned in namespace
[It] should support (non-root,0777,tmpfs) [LinuxOnly] [NodeConformance] [Conformance]
  /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:630
STEP: Creating a pod to test emptydir 0777 on tmpfs
Jul 18 19:53:31.395: INFO: Waiting up to 5m0s for pod "pod-2fdeaa51-6587-40ca-961f-25164f17c007" in namespace "emptydir-1070" to be "Succeeded or Failed"
Jul 18 19:53:31.426: INFO: Pod "pod-2fdeaa51-6587-40ca-961f-25164f17c007": Phase="Pending", Reason="", readiness=false. Elapsed: 30.953127ms
Jul 18 19:53:33.457: INFO: Pod "pod-2fdeaa51-6587-40ca-961f-25164f17c007": Phase="Pending", Reason="", readiness=false. Elapsed: 2.062426705s
Jul 18 19:53:35.490: INFO: Pod "pod-2fdeaa51-6587-40ca-961f-25164f17c007": Phase="Pending", Reason="", readiness=false. Elapsed: 4.095708365s
Jul 18 19:53:37.523: INFO: Pod "pod-2fdeaa51-6587-40ca-961f-25164f17c007": Phase="Pending", Reason="", readiness=false. Elapsed: 6.128229077s
Jul 18 19:53:39.556: INFO: Pod "pod-2fdeaa51-6587-40ca-961f-25164f17c007": Phase="Pending", Reason="", readiness=false. Elapsed: 8.161090461s
Jul 18 19:53:41.588: INFO: Pod "pod-2fdeaa51-6587-40ca-961f-25164f17c007": Phase="Pending", Reason="", readiness=false. Elapsed: 10.193602081s
Jul 18 19:53:43.621: INFO: Pod "pod-2fdeaa51-6587-40ca-961f-25164f17c007": Phase="Succeeded", Reason="", readiness=false. Elapsed: 12.226709456s
STEP: Saw pod success
Jul 18 19:53:43.622: INFO: Pod "pod-2fdeaa51-6587-40ca-961f-25164f17c007" satisfied condition "Succeeded or Failed"
Jul 18 19:53:43.655: INFO: Trying to get logs from node 10.116.0.5 pod pod-2fdeaa51-6587-40ca-961f-25164f17c007 container test-container: <nil>
STEP: delete the pod
Jul 18 19:53:43.733: INFO: Waiting for pod pod-2fdeaa51-6587-40ca-961f-25164f17c007 to disappear
Jul 18 19:53:43.764: INFO: Pod pod-2fdeaa51-6587-40ca-961f-25164f17c007 no longer exists
[AfterEach] [sig-storage] EmptyDir volumes
  /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:186
... skipping 4 lines ...
• [SLOW TEST:12.661 seconds]
[sig-storage] EmptyDir volumes
/workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/common/storage/framework.go:23
  should support (non-root,0777,tmpfs) [LinuxOnly] [NodeConformance] [Conformance]
  /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:630
------------------------------
{"msg":"PASSED [sig-storage] EmptyDir volumes should support (non-root,0777,tmpfs) [LinuxOnly] [NodeConformance] [Conformance]","total":-1,"completed":10,"skipped":54,"failed":0}

SS
------------------------------
[BeforeEach] [sig-network] Networking
  /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:185
STEP: Creating a kubernetes client
... skipping 74 lines ...
Jul 18 19:53:21.366: INFO: PersistentVolumeClaim pvc-57kxq found but phase is Pending instead of Bound.
Jul 18 19:53:23.401: INFO: PersistentVolumeClaim pvc-57kxq found and phase=Bound (2.065838309s)
Jul 18 19:53:23.401: INFO: Waiting up to 3m0s for PersistentVolume local-pcnlz to have phase Bound
Jul 18 19:53:23.432: INFO: PersistentVolume local-pcnlz found and phase=Bound (30.60732ms)
STEP: Creating pod pod-subpath-test-preprovisionedpv-4hmd
STEP: Creating a pod to test subpath
Jul 18 19:53:23.529: INFO: Waiting up to 5m0s for pod "pod-subpath-test-preprovisionedpv-4hmd" in namespace "provisioning-8275" to be "Succeeded or Failed"
Jul 18 19:53:23.560: INFO: Pod "pod-subpath-test-preprovisionedpv-4hmd": Phase="Pending", Reason="", readiness=false. Elapsed: 30.577602ms
Jul 18 19:53:25.592: INFO: Pod "pod-subpath-test-preprovisionedpv-4hmd": Phase="Pending", Reason="", readiness=false. Elapsed: 2.06302286s
Jul 18 19:53:27.624: INFO: Pod "pod-subpath-test-preprovisionedpv-4hmd": Phase="Pending", Reason="", readiness=false. Elapsed: 4.095087887s
Jul 18 19:53:29.658: INFO: Pod "pod-subpath-test-preprovisionedpv-4hmd": Phase="Pending", Reason="", readiness=false. Elapsed: 6.128922739s
Jul 18 19:53:31.691: INFO: Pod "pod-subpath-test-preprovisionedpv-4hmd": Phase="Pending", Reason="", readiness=false. Elapsed: 8.161821812s
Jul 18 19:53:33.724: INFO: Pod "pod-subpath-test-preprovisionedpv-4hmd": Phase="Pending", Reason="", readiness=false. Elapsed: 10.195018439s
Jul 18 19:53:35.996: INFO: Pod "pod-subpath-test-preprovisionedpv-4hmd": Phase="Pending", Reason="", readiness=false. Elapsed: 12.466549083s
Jul 18 19:53:38.028: INFO: Pod "pod-subpath-test-preprovisionedpv-4hmd": Phase="Pending", Reason="", readiness=false. Elapsed: 14.498948912s
Jul 18 19:53:40.061: INFO: Pod "pod-subpath-test-preprovisionedpv-4hmd": Phase="Pending", Reason="", readiness=false. Elapsed: 16.53162195s
Jul 18 19:53:42.094: INFO: Pod "pod-subpath-test-preprovisionedpv-4hmd": Phase="Pending", Reason="", readiness=false. Elapsed: 18.564554359s
Jul 18 19:53:44.127: INFO: Pod "pod-subpath-test-preprovisionedpv-4hmd": Phase="Succeeded", Reason="", readiness=false. Elapsed: 20.597332675s
STEP: Saw pod success
Jul 18 19:53:44.127: INFO: Pod "pod-subpath-test-preprovisionedpv-4hmd" satisfied condition "Succeeded or Failed"
Jul 18 19:53:44.158: INFO: Trying to get logs from node 10.116.0.7 pod pod-subpath-test-preprovisionedpv-4hmd container test-container-subpath-preprovisionedpv-4hmd: <nil>
STEP: delete the pod
Jul 18 19:53:44.246: INFO: Waiting for pod pod-subpath-test-preprovisionedpv-4hmd to disappear
Jul 18 19:53:44.277: INFO: Pod pod-subpath-test-preprovisionedpv-4hmd no longer exists
STEP: Deleting pod pod-subpath-test-preprovisionedpv-4hmd
Jul 18 19:53:44.277: INFO: Deleting pod "pod-subpath-test-preprovisionedpv-4hmd" in namespace "provisioning-8275"
... skipping 24 lines ...
  /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/storage/in_tree_volumes.go:58
    [Testpattern: Pre-provisioned PV (default fs)] subPath
    /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/storage/framework/testsuite.go:50
      should support existing single file [LinuxOnly]
      /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/storage/testsuites/subpath.go:219
------------------------------
{"msg":"PASSED [sig-storage] In-tree Volumes [Driver: local][LocalVolumeType: block] [Testpattern: Pre-provisioned PV (default fs)] subPath should support existing single file [LinuxOnly]","total":-1,"completed":6,"skipped":30,"failed":0}
[BeforeEach] [Testpattern: Dynamic PV (default fs)] subPath
  /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/storage/framework/testsuite.go:51
Jul 18 19:53:45.346: INFO: Driver local doesn't support DynamicPV -- skipping
[AfterEach] [Testpattern: Dynamic PV (default fs)] subPath
  /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:186

... skipping 83 lines ...
  /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/storage/in_tree_volumes.go:58
    [Testpattern: Pre-provisioned PV (block volmode)] volumeMode
    /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/storage/framework/testsuite.go:50
      should not mount / map unused volumes in a pod [LinuxOnly]
      /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/storage/testsuites/volumemode.go:351
------------------------------
{"msg":"PASSED [sig-storage] In-tree Volumes [Driver: local][LocalVolumeType: block] [Testpattern: Pre-provisioned PV (block volmode)] volumeMode should not mount / map unused volumes in a pod [LinuxOnly]","total":-1,"completed":6,"skipped":46,"failed":0}

SSSSSSSSSSS
------------------------------
[BeforeEach] [Testpattern: Pre-provisioned PV (default fs)] subPath
  /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/storage/framework/testsuite.go:51
Jul 18 19:53:45.986: INFO: Driver csi-hostpath doesn't support PreprovisionedPV -- skipping
... skipping 143 lines ...
Jul 18 19:52:41.795: INFO: PersistentVolumeClaim csi-hostpathbc9d4 found but phase is Pending instead of Bound.
Jul 18 19:52:43.825: INFO: PersistentVolumeClaim csi-hostpathbc9d4 found but phase is Pending instead of Bound.
Jul 18 19:52:45.856: INFO: PersistentVolumeClaim csi-hostpathbc9d4 found but phase is Pending instead of Bound.
Jul 18 19:52:47.886: INFO: PersistentVolumeClaim csi-hostpathbc9d4 found and phase=Bound (8.153149055s)
STEP: Creating pod pod-subpath-test-dynamicpv-h979
STEP: Creating a pod to test subpath
Jul 18 19:52:47.983: INFO: Waiting up to 5m0s for pod "pod-subpath-test-dynamicpv-h979" in namespace "provisioning-5461" to be "Succeeded or Failed"
Jul 18 19:52:48.014: INFO: Pod "pod-subpath-test-dynamicpv-h979": Phase="Pending", Reason="", readiness=false. Elapsed: 30.94229ms
Jul 18 19:52:50.046: INFO: Pod "pod-subpath-test-dynamicpv-h979": Phase="Pending", Reason="", readiness=false. Elapsed: 2.06275839s
Jul 18 19:52:52.079: INFO: Pod "pod-subpath-test-dynamicpv-h979": Phase="Pending", Reason="", readiness=false. Elapsed: 4.095681839s
Jul 18 19:52:54.110: INFO: Pod "pod-subpath-test-dynamicpv-h979": Phase="Pending", Reason="", readiness=false. Elapsed: 6.127156752s
Jul 18 19:52:56.141: INFO: Pod "pod-subpath-test-dynamicpv-h979": Phase="Pending", Reason="", readiness=false. Elapsed: 8.158415395s
Jul 18 19:52:58.172: INFO: Pod "pod-subpath-test-dynamicpv-h979": Phase="Pending", Reason="", readiness=false. Elapsed: 10.189360878s
... skipping 3 lines ...
Jul 18 19:53:06.298: INFO: Pod "pod-subpath-test-dynamicpv-h979": Phase="Pending", Reason="", readiness=false. Elapsed: 18.314714767s
Jul 18 19:53:08.331: INFO: Pod "pod-subpath-test-dynamicpv-h979": Phase="Pending", Reason="", readiness=false. Elapsed: 20.347825701s
Jul 18 19:53:10.362: INFO: Pod "pod-subpath-test-dynamicpv-h979": Phase="Pending", Reason="", readiness=false. Elapsed: 22.378857618s
Jul 18 19:53:12.394: INFO: Pod "pod-subpath-test-dynamicpv-h979": Phase="Pending", Reason="", readiness=false. Elapsed: 24.410871336s
Jul 18 19:53:14.449: INFO: Pod "pod-subpath-test-dynamicpv-h979": Phase="Succeeded", Reason="", readiness=false. Elapsed: 26.466106793s
STEP: Saw pod success
Jul 18 19:53:14.449: INFO: Pod "pod-subpath-test-dynamicpv-h979" satisfied condition "Succeeded or Failed"
Jul 18 19:53:14.480: INFO: Trying to get logs from node 10.116.0.5 pod pod-subpath-test-dynamicpv-h979 container test-container-subpath-dynamicpv-h979: <nil>
STEP: delete the pod
Jul 18 19:53:14.557: INFO: Waiting for pod pod-subpath-test-dynamicpv-h979 to disappear
Jul 18 19:53:14.586: INFO: Pod pod-subpath-test-dynamicpv-h979 no longer exists
STEP: Deleting pod pod-subpath-test-dynamicpv-h979
Jul 18 19:53:14.586: INFO: Deleting pod "pod-subpath-test-dynamicpv-h979" in namespace "provisioning-5461"
... skipping 54 lines ...
  /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/storage/csi_volumes.go:40
    [Testpattern: Dynamic PV (default fs)] subPath
    /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/storage/framework/testsuite.go:50
      should support readOnly file specified in the volumeMount [LinuxOnly]
      /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/storage/testsuites/subpath.go:379
------------------------------
{"msg":"PASSED [sig-storage] CSI Volumes [Driver: csi-hostpath] [Testpattern: Dynamic PV (default fs)] subPath should support readOnly file specified in the volumeMount [LinuxOnly]","total":-1,"completed":3,"skipped":49,"failed":0}

SS
------------------------------
[BeforeEach] [Testpattern: Dynamic PV (default fs)] provisioning
  /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/storage/framework/testsuite.go:51
Jul 18 19:53:51.174: INFO: Driver local doesn't support DynamicPV -- skipping
... skipping 41 lines ...
Jul 18 19:53:20.174: INFO: PersistentVolumeClaim pvc-7tphc found but phase is Pending instead of Bound.
Jul 18 19:53:22.205: INFO: PersistentVolumeClaim pvc-7tphc found and phase=Bound (8.16546578s)
Jul 18 19:53:22.205: INFO: Waiting up to 3m0s for PersistentVolume local-rjmdf to have phase Bound
Jul 18 19:53:22.236: INFO: PersistentVolume local-rjmdf found and phase=Bound (30.427403ms)
STEP: Creating pod pod-subpath-test-preprovisionedpv-jws6
STEP: Creating a pod to test subpath
Jul 18 19:53:22.335: INFO: Waiting up to 5m0s for pod "pod-subpath-test-preprovisionedpv-jws6" in namespace "provisioning-9616" to be "Succeeded or Failed"
Jul 18 19:53:22.366: INFO: Pod "pod-subpath-test-preprovisionedpv-jws6": Phase="Pending", Reason="", readiness=false. Elapsed: 31.314248ms
Jul 18 19:53:24.399: INFO: Pod "pod-subpath-test-preprovisionedpv-jws6": Phase="Pending", Reason="", readiness=false. Elapsed: 2.063847125s
Jul 18 19:53:26.431: INFO: Pod "pod-subpath-test-preprovisionedpv-jws6": Phase="Pending", Reason="", readiness=false. Elapsed: 4.095828243s
Jul 18 19:53:28.464: INFO: Pod "pod-subpath-test-preprovisionedpv-jws6": Phase="Pending", Reason="", readiness=false. Elapsed: 6.12920408s
Jul 18 19:53:30.496: INFO: Pod "pod-subpath-test-preprovisionedpv-jws6": Phase="Pending", Reason="", readiness=false. Elapsed: 8.160576683s
Jul 18 19:53:32.528: INFO: Pod "pod-subpath-test-preprovisionedpv-jws6": Phase="Pending", Reason="", readiness=false. Elapsed: 10.192864801s
Jul 18 19:53:34.561: INFO: Pod "pod-subpath-test-preprovisionedpv-jws6": Phase="Pending", Reason="", readiness=false. Elapsed: 12.225555488s
Jul 18 19:53:36.597: INFO: Pod "pod-subpath-test-preprovisionedpv-jws6": Phase="Pending", Reason="", readiness=false. Elapsed: 14.262425736s
Jul 18 19:53:38.633: INFO: Pod "pod-subpath-test-preprovisionedpv-jws6": Phase="Pending", Reason="", readiness=false. Elapsed: 16.297669726s
Jul 18 19:53:40.667: INFO: Pod "pod-subpath-test-preprovisionedpv-jws6": Phase="Succeeded", Reason="", readiness=false. Elapsed: 18.332362246s
STEP: Saw pod success
Jul 18 19:53:40.667: INFO: Pod "pod-subpath-test-preprovisionedpv-jws6" satisfied condition "Succeeded or Failed"
Jul 18 19:53:40.699: INFO: Trying to get logs from node 10.116.0.7 pod pod-subpath-test-preprovisionedpv-jws6 container test-container-subpath-preprovisionedpv-jws6: <nil>
STEP: delete the pod
Jul 18 19:53:40.781: INFO: Waiting for pod pod-subpath-test-preprovisionedpv-jws6 to disappear
Jul 18 19:53:40.812: INFO: Pod pod-subpath-test-preprovisionedpv-jws6 no longer exists
STEP: Deleting pod pod-subpath-test-preprovisionedpv-jws6
Jul 18 19:53:40.812: INFO: Deleting pod "pod-subpath-test-preprovisionedpv-jws6" in namespace "provisioning-9616"
STEP: Creating pod pod-subpath-test-preprovisionedpv-jws6
STEP: Creating a pod to test subpath
Jul 18 19:53:40.893: INFO: Waiting up to 5m0s for pod "pod-subpath-test-preprovisionedpv-jws6" in namespace "provisioning-9616" to be "Succeeded or Failed"
Jul 18 19:53:40.925: INFO: Pod "pod-subpath-test-preprovisionedpv-jws6": Phase="Pending", Reason="", readiness=false. Elapsed: 31.56341ms
Jul 18 19:53:42.958: INFO: Pod "pod-subpath-test-preprovisionedpv-jws6": Phase="Pending", Reason="", readiness=false. Elapsed: 2.064150091s
Jul 18 19:53:44.990: INFO: Pod "pod-subpath-test-preprovisionedpv-jws6": Phase="Pending", Reason="", readiness=false. Elapsed: 4.096793431s
Jul 18 19:53:47.024: INFO: Pod "pod-subpath-test-preprovisionedpv-jws6": Phase="Pending", Reason="", readiness=false. Elapsed: 6.130031862s
Jul 18 19:53:49.056: INFO: Pod "pod-subpath-test-preprovisionedpv-jws6": Phase="Pending", Reason="", readiness=false. Elapsed: 8.162816752s
Jul 18 19:53:51.092: INFO: Pod "pod-subpath-test-preprovisionedpv-jws6": Phase="Succeeded", Reason="", readiness=false. Elapsed: 10.197967094s
STEP: Saw pod success
Jul 18 19:53:51.092: INFO: Pod "pod-subpath-test-preprovisionedpv-jws6" satisfied condition "Succeeded or Failed"
Jul 18 19:53:51.123: INFO: Trying to get logs from node 10.116.0.7 pod pod-subpath-test-preprovisionedpv-jws6 container test-container-subpath-preprovisionedpv-jws6: <nil>
STEP: delete the pod
Jul 18 19:53:51.207: INFO: Waiting for pod pod-subpath-test-preprovisionedpv-jws6 to disappear
Jul 18 19:53:51.237: INFO: Pod pod-subpath-test-preprovisionedpv-jws6 no longer exists
STEP: Deleting pod pod-subpath-test-preprovisionedpv-jws6
Jul 18 19:53:51.237: INFO: Deleting pod "pod-subpath-test-preprovisionedpv-jws6" in namespace "provisioning-9616"
... skipping 19 lines ...
  /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/storage/in_tree_volumes.go:58
    [Testpattern: Pre-provisioned PV (default fs)] subPath
    /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/storage/framework/testsuite.go:50
      should support existing directories when readOnly specified in the volumeSource
      /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/storage/testsuites/subpath.go:394
------------------------------
{"msg":"PASSED [sig-storage] In-tree Volumes [Driver: local][LocalVolumeType: dir-link] [Testpattern: Pre-provisioned PV (default fs)] subPath should support existing directories when readOnly specified in the volumeSource","total":-1,"completed":8,"skipped":61,"failed":0}
[BeforeEach] [Testpattern: Dynamic PV (default fs)(allowExpansion)] volume-expand
  /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/storage/framework/testsuite.go:51
Jul 18 19:53:51.766: INFO: Driver local doesn't support DynamicPV -- skipping
[AfterEach] [Testpattern: Dynamic PV (default fs)(allowExpansion)] volume-expand
  /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:186

... skipping 52 lines ...
  /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:186
Jul 18 19:53:52.188: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready
STEP: Destroying namespace "kubectl-5875" for this suite.

•
------------------------------
{"msg":"PASSED [sig-cli] Kubectl client Kubectl version should check is all data is printed  [Conformance]","total":-1,"completed":9,"skipped":68,"failed":0}

SSSSS
------------------------------
[BeforeEach] [Testpattern: Pre-provisioned PV (default fs)] subPath
  /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/storage/framework/testsuite.go:51
Jul 18 19:53:52.311: INFO: Only supported for providers [gce gke] (not skeleton)
... skipping 62 lines ...
Jul 18 19:52:38.374: INFO: creating *v1.CSIDriver: csi-mock-csi-mock-volumes-5951
Jul 18 19:52:38.412: INFO: creating *v1.StatefulSet: csi-mock-volumes-5951-8767/csi-mockplugin-attacher
Jul 18 19:52:38.445: INFO: waiting up to 4m0s for CSIDriver "csi-mock-csi-mock-volumes-5951"
Jul 18 19:52:38.476: INFO: waiting for CSIDriver csi-mock-csi-mock-volumes-5951 to register on node 10.116.0.7
STEP: Creating pod
STEP: checking for CSIInlineVolumes feature
Jul 18 19:53:06.403: INFO: Error getting logs for pod inline-volume-7nlpx: the server rejected our request for an unknown reason (get pods inline-volume-7nlpx)
Jul 18 19:53:06.437: INFO: Deleting pod "inline-volume-7nlpx" in namespace "csi-mock-volumes-5951"
Jul 18 19:53:06.471: INFO: Wait up to 5m0s for pod "inline-volume-7nlpx" to be fully deleted
STEP: Deleting the previously created pod
Jul 18 19:53:22.535: INFO: Deleting pod "pvc-volume-tester-lr4g6" in namespace "csi-mock-volumes-5951"
Jul 18 19:53:22.572: INFO: Wait up to 5m0s for pod "pvc-volume-tester-lr4g6" to be fully deleted
STEP: Checking CSI driver logs
Jul 18 19:53:36.677: INFO: Found volume attribute csi.storage.k8s.io/pod.name: pvc-volume-tester-lr4g6
Jul 18 19:53:36.677: INFO: Found volume attribute csi.storage.k8s.io/pod.namespace: csi-mock-volumes-5951
Jul 18 19:53:36.677: INFO: Found volume attribute csi.storage.k8s.io/pod.uid: 58d049f8-9850-48ff-973b-ea50efb1740f
Jul 18 19:53:36.677: INFO: Found volume attribute csi.storage.k8s.io/serviceAccount.name: default
Jul 18 19:53:36.677: INFO: Found volume attribute csi.storage.k8s.io/ephemeral: true
Jul 18 19:53:36.677: INFO: Found NodeUnpublishVolume: {json: {"Method":"/csi.v1.Node/NodeUnpublishVolume","Request":{"volume_id":"csi-9093b7beddc79dc70828934c049e7166e9c479f17d6cbc49530fc62f41b20018","target_path":"/var/lib/kubelet/pods/58d049f8-9850-48ff-973b-ea50efb1740f/volumes/kubernetes.io~csi/my-volume/mount"},"Response":{},"Error":"","FullError":null} Method:NodeUnpublishVolume Request:{VolumeContext:map[]} FullError:{Code:OK Message:} Error:}
STEP: Deleting pod pvc-volume-tester-lr4g6
Jul 18 19:53:36.677: INFO: Deleting pod "pvc-volume-tester-lr4g6" in namespace "csi-mock-volumes-5951"
STEP: Cleaning up resources
STEP: deleting the test namespace: csi-mock-volumes-5951
STEP: Waiting for namespaces [csi-mock-volumes-5951] to vanish
STEP: uninstalling csi mock driver
... skipping 40 lines ...
/workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/storage/utils/framework.go:23
  CSI workload information using mock driver
  /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/storage/csi_mock_volume.go:443
    contain ephemeral=true when using inline volume
    /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/storage/csi_mock_volume.go:493
------------------------------
{"msg":"PASSED [sig-storage] CSI mock volume CSI workload information using mock driver contain ephemeral=true when using inline volume","total":-1,"completed":5,"skipped":37,"failed":0}

S
------------------------------
[BeforeEach] [Testpattern: Inline-volume (ext3)] volumes
  /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/storage/framework/testsuite.go:51
Jul 18 19:54:00.137: INFO: Driver local doesn't support InlineVolume -- skipping
... skipping 86 lines ...
Jul 18 19:53:20.860: INFO: PersistentVolumeClaim pvc-wtrhv found but phase is Pending instead of Bound.
Jul 18 19:53:22.892: INFO: PersistentVolumeClaim pvc-wtrhv found and phase=Bound (10.190817949s)
Jul 18 19:53:22.892: INFO: Waiting up to 3m0s for PersistentVolume local-7p798 to have phase Bound
Jul 18 19:53:22.924: INFO: PersistentVolume local-7p798 found and phase=Bound (32.670821ms)
STEP: Creating pod pod-subpath-test-preprovisionedpv-rdk7
STEP: Creating a pod to test subpath
Jul 18 19:53:23.022: INFO: Waiting up to 5m0s for pod "pod-subpath-test-preprovisionedpv-rdk7" in namespace "provisioning-4501" to be "Succeeded or Failed"
Jul 18 19:53:23.054: INFO: Pod "pod-subpath-test-preprovisionedpv-rdk7": Phase="Pending", Reason="", readiness=false. Elapsed: 32.139891ms
Jul 18 19:53:25.086: INFO: Pod "pod-subpath-test-preprovisionedpv-rdk7": Phase="Pending", Reason="", readiness=false. Elapsed: 2.063963228s
Jul 18 19:53:27.118: INFO: Pod "pod-subpath-test-preprovisionedpv-rdk7": Phase="Pending", Reason="", readiness=false. Elapsed: 4.096420257s
Jul 18 19:53:29.150: INFO: Pod "pod-subpath-test-preprovisionedpv-rdk7": Phase="Pending", Reason="", readiness=false. Elapsed: 6.128598735s
Jul 18 19:53:31.185: INFO: Pod "pod-subpath-test-preprovisionedpv-rdk7": Phase="Pending", Reason="", readiness=false. Elapsed: 8.162918694s
Jul 18 19:53:33.218: INFO: Pod "pod-subpath-test-preprovisionedpv-rdk7": Phase="Pending", Reason="", readiness=false. Elapsed: 10.195652625s
... skipping 2 lines ...
Jul 18 19:53:39.315: INFO: Pod "pod-subpath-test-preprovisionedpv-rdk7": Phase="Pending", Reason="", readiness=false. Elapsed: 16.293316152s
Jul 18 19:53:41.349: INFO: Pod "pod-subpath-test-preprovisionedpv-rdk7": Phase="Pending", Reason="", readiness=false. Elapsed: 18.327369882s
Jul 18 19:53:43.386: INFO: Pod "pod-subpath-test-preprovisionedpv-rdk7": Phase="Pending", Reason="", readiness=false. Elapsed: 20.364012448s
Jul 18 19:53:45.418: INFO: Pod "pod-subpath-test-preprovisionedpv-rdk7": Phase="Pending", Reason="", readiness=false. Elapsed: 22.395715343s
Jul 18 19:53:47.450: INFO: Pod "pod-subpath-test-preprovisionedpv-rdk7": Phase="Succeeded", Reason="", readiness=false. Elapsed: 24.428594321s
STEP: Saw pod success
Jul 18 19:53:47.451: INFO: Pod "pod-subpath-test-preprovisionedpv-rdk7" satisfied condition "Succeeded or Failed"
Jul 18 19:53:47.482: INFO: Trying to get logs from node 10.116.0.7 pod pod-subpath-test-preprovisionedpv-rdk7 container test-container-subpath-preprovisionedpv-rdk7: <nil>
STEP: delete the pod
Jul 18 19:53:47.560: INFO: Waiting for pod pod-subpath-test-preprovisionedpv-rdk7 to disappear
Jul 18 19:53:47.595: INFO: Pod pod-subpath-test-preprovisionedpv-rdk7 no longer exists
STEP: Deleting pod pod-subpath-test-preprovisionedpv-rdk7
Jul 18 19:53:47.595: INFO: Deleting pod "pod-subpath-test-preprovisionedpv-rdk7" in namespace "provisioning-4501"
STEP: Creating pod pod-subpath-test-preprovisionedpv-rdk7
STEP: Creating a pod to test subpath
Jul 18 19:53:47.669: INFO: Waiting up to 5m0s for pod "pod-subpath-test-preprovisionedpv-rdk7" in namespace "provisioning-4501" to be "Succeeded or Failed"
Jul 18 19:53:47.702: INFO: Pod "pod-subpath-test-preprovisionedpv-rdk7": Phase="Pending", Reason="", readiness=false. Elapsed: 33.134833ms
Jul 18 19:53:49.734: INFO: Pod "pod-subpath-test-preprovisionedpv-rdk7": Phase="Pending", Reason="", readiness=false. Elapsed: 2.065383413s
Jul 18 19:53:51.772: INFO: Pod "pod-subpath-test-preprovisionedpv-rdk7": Phase="Pending", Reason="", readiness=false. Elapsed: 4.102774319s
Jul 18 19:53:53.805: INFO: Pod "pod-subpath-test-preprovisionedpv-rdk7": Phase="Pending", Reason="", readiness=false. Elapsed: 6.136257821s
Jul 18 19:53:55.837: INFO: Pod "pod-subpath-test-preprovisionedpv-rdk7": Phase="Pending", Reason="", readiness=false. Elapsed: 8.168489396s
Jul 18 19:53:57.869: INFO: Pod "pod-subpath-test-preprovisionedpv-rdk7": Phase="Pending", Reason="", readiness=false. Elapsed: 10.200261144s
Jul 18 19:53:59.903: INFO: Pod "pod-subpath-test-preprovisionedpv-rdk7": Phase="Succeeded", Reason="", readiness=false. Elapsed: 12.234621203s
STEP: Saw pod success
Jul 18 19:53:59.904: INFO: Pod "pod-subpath-test-preprovisionedpv-rdk7" satisfied condition "Succeeded or Failed"
Jul 18 19:53:59.935: INFO: Trying to get logs from node 10.116.0.7 pod pod-subpath-test-preprovisionedpv-rdk7 container test-container-subpath-preprovisionedpv-rdk7: <nil>
STEP: delete the pod
Jul 18 19:54:00.019: INFO: Waiting for pod pod-subpath-test-preprovisionedpv-rdk7 to disappear
Jul 18 19:54:00.054: INFO: Pod pod-subpath-test-preprovisionedpv-rdk7 no longer exists
STEP: Deleting pod pod-subpath-test-preprovisionedpv-rdk7
Jul 18 19:54:00.054: INFO: Deleting pod "pod-subpath-test-preprovisionedpv-rdk7" in namespace "provisioning-4501"
... skipping 19 lines ...
  /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/storage/in_tree_volumes.go:58
    [Testpattern: Pre-provisioned PV (default fs)] subPath
    /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/storage/framework/testsuite.go:50
      should support existing directories when readOnly specified in the volumeSource
      /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/storage/testsuites/subpath.go:394
------------------------------
{"msg":"PASSED [sig-storage] In-tree Volumes [Driver: local][LocalVolumeType: dir-link-bindmounted] [Testpattern: Pre-provisioned PV (default fs)] subPath should support existing directories when readOnly specified in the volumeSource","total":-1,"completed":6,"skipped":61,"failed":0}
[BeforeEach] [Testpattern: Dynamic PV (default fs)] fsgroupchangepolicy
  /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/storage/framework/testsuite.go:51
Jul 18 19:54:00.572: INFO: Driver local doesn't support DynamicPV -- skipping
[AfterEach] [Testpattern: Dynamic PV (default fs)] fsgroupchangepolicy
  /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:186

... skipping 11 lines ...
      Driver local doesn't support DynamicPV -- skipping

      /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/storage/framework/testsuite.go:116
------------------------------
SSSS
------------------------------
{"msg":"PASSED [sig-apps] ReplicaSet Replicaset should have a working scale subresource [Conformance]","total":-1,"completed":7,"skipped":15,"failed":0}
[BeforeEach] [sig-storage] Ephemeralstorage
  /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:185
STEP: Creating a kubernetes client
Jul 18 19:53:16.367: INFO: >>> kubeConfig: /root/.kube/config
STEP: Building a namespace api object, basename pv
STEP: Waiting for a default service account to be provisioned in namespace
... skipping 15 lines ...
/workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/storage/utils/framework.go:23
  When pod refers to non-existent ephemeral storage
  /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/storage/ephemeral_volume.go:53
    should allow deletion of pod with invalid volume : projected
    /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/storage/ephemeral_volume.go:55
------------------------------
{"msg":"PASSED [sig-storage] Ephemeralstorage When pod refers to non-existent ephemeral storage should allow deletion of pod with invalid volume : projected","total":-1,"completed":8,"skipped":15,"failed":0}

SSS
------------------------------
[BeforeEach] [Testpattern: Dynamic PV (default fs)(allowExpansion)] volume-expand
  /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/storage/framework/testsuite.go:51
Jul 18 19:54:04.783: INFO: Driver local doesn't support DynamicPV -- skipping
... skipping 41 lines ...
  /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/common/node/runtime.go:41
    when running a container with a new image
    /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/common/node/runtime.go:266
      should be able to pull image [NodeConformance]
      /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/common/node/runtime.go:382
------------------------------
{"msg":"PASSED [sig-node] Container Runtime blackbox test when running a container with a new image should be able to pull image [NodeConformance]","total":-1,"completed":7,"skipped":66,"failed":0}
[BeforeEach] [Testpattern: Pre-provisioned PV (block volmode)] volumes
  /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/storage/framework/testsuite.go:51
Jul 18 19:54:05.006: INFO: Driver emptydir doesn't support PreprovisionedPV -- skipping
[AfterEach] [Testpattern: Pre-provisioned PV (block volmode)] volumes
  /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:186

... skipping 63 lines ...
STEP: Building a namespace api object, basename projected
STEP: Waiting for a default service account to be provisioned in namespace
[It] should be consumable in multiple volumes in the same pod [NodeConformance] [Conformance]
  /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:630
STEP: Creating configMap with name projected-configmap-test-volume-4d9e9386-1f0d-4a85-ac26-8fa7f8b375ee
STEP: Creating a pod to test consume configMaps
Jul 18 19:53:45.047: INFO: Waiting up to 5m0s for pod "pod-projected-configmaps-90a0df76-746f-4747-97ab-16f306bed90c" in namespace "projected-1510" to be "Succeeded or Failed"
Jul 18 19:53:45.077: INFO: Pod "pod-projected-configmaps-90a0df76-746f-4747-97ab-16f306bed90c": Phase="Pending", Reason="", readiness=false. Elapsed: 30.302183ms
Jul 18 19:53:47.108: INFO: Pod "pod-projected-configmaps-90a0df76-746f-4747-97ab-16f306bed90c": Phase="Pending", Reason="", readiness=false. Elapsed: 2.061516138s
Jul 18 19:53:49.140: INFO: Pod "pod-projected-configmaps-90a0df76-746f-4747-97ab-16f306bed90c": Phase="Pending", Reason="", readiness=false. Elapsed: 4.09294556s
Jul 18 19:53:51.171: INFO: Pod "pod-projected-configmaps-90a0df76-746f-4747-97ab-16f306bed90c": Phase="Pending", Reason="", readiness=false. Elapsed: 6.1235496s
Jul 18 19:53:53.206: INFO: Pod "pod-projected-configmaps-90a0df76-746f-4747-97ab-16f306bed90c": Phase="Pending", Reason="", readiness=false. Elapsed: 8.158805645s
Jul 18 19:53:55.237: INFO: Pod "pod-projected-configmaps-90a0df76-746f-4747-97ab-16f306bed90c": Phase="Pending", Reason="", readiness=false. Elapsed: 10.190425596s
Jul 18 19:53:57.270: INFO: Pod "pod-projected-configmaps-90a0df76-746f-4747-97ab-16f306bed90c": Phase="Pending", Reason="", readiness=false. Elapsed: 12.222947762s
Jul 18 19:53:59.304: INFO: Pod "pod-projected-configmaps-90a0df76-746f-4747-97ab-16f306bed90c": Phase="Pending", Reason="", readiness=false. Elapsed: 14.256796722s
Jul 18 19:54:01.336: INFO: Pod "pod-projected-configmaps-90a0df76-746f-4747-97ab-16f306bed90c": Phase="Pending", Reason="", readiness=false. Elapsed: 16.288880457s
Jul 18 19:54:03.367: INFO: Pod "pod-projected-configmaps-90a0df76-746f-4747-97ab-16f306bed90c": Phase="Pending", Reason="", readiness=false. Elapsed: 18.320101589s
Jul 18 19:54:05.398: INFO: Pod "pod-projected-configmaps-90a0df76-746f-4747-97ab-16f306bed90c": Phase="Succeeded", Reason="", readiness=false. Elapsed: 20.351510314s
STEP: Saw pod success
Jul 18 19:54:05.399: INFO: Pod "pod-projected-configmaps-90a0df76-746f-4747-97ab-16f306bed90c" satisfied condition "Succeeded or Failed"
Jul 18 19:54:05.429: INFO: Trying to get logs from node 10.116.0.5 pod pod-projected-configmaps-90a0df76-746f-4747-97ab-16f306bed90c container projected-configmap-volume-test: <nil>
STEP: delete the pod
Jul 18 19:54:05.503: INFO: Waiting for pod pod-projected-configmaps-90a0df76-746f-4747-97ab-16f306bed90c to disappear
Jul 18 19:54:05.534: INFO: Pod pod-projected-configmaps-90a0df76-746f-4747-97ab-16f306bed90c no longer exists
[AfterEach] [sig-storage] Projected configMap
  /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:186
... skipping 34 lines ...
• [SLOW TEST:35.327 seconds]
[sig-api-machinery] ResourceQuota
/workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/apimachinery/framework.go:23
  should create a ResourceQuota and capture the life of a custom resource.
  /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/apimachinery/resource_quota.go:583
------------------------------
{"msg":"PASSED [sig-api-machinery] ResourceQuota should create a ResourceQuota and capture the life of a custom resource.","total":-1,"completed":7,"skipped":26,"failed":0}
[BeforeEach] [Testpattern: Pre-provisioned PV (ext4)] volumes
  /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/storage/framework/testsuite.go:51
Jul 18 19:54:06.667: INFO: Driver local doesn't support ext4 -- skipping
[AfterEach] [Testpattern: Pre-provisioned PV (ext4)] volumes
  /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:186

... skipping 138 lines ...
STEP: creating an object not containing a namespace with in-cluster config
Jul 18 19:54:05.765: INFO: Running '/tmp/kubectl2220469345/kubectl --server=https://143.244.215.182 --kubeconfig=/root/.kube/config --namespace=kubectl-4169 exec httpd -- /bin/sh -x -c /tmp/kubectl create -f /tmp/invalid-configmap-without-namespace.yaml --v=6 2>&1'
Jul 18 19:54:06.402: INFO: rc: 255
STEP: trying to use kubectl with invalid token
Jul 18 19:54:06.402: INFO: Running '/tmp/kubectl2220469345/kubectl --server=https://143.244.215.182 --kubeconfig=/root/.kube/config --namespace=kubectl-4169 exec httpd -- /bin/sh -x -c /tmp/kubectl get pods --token=invalid --v=7 2>&1'
Jul 18 19:54:06.881: INFO: rc: 255
Jul 18 19:54:06.881: INFO: got err error running /tmp/kubectl2220469345/kubectl --server=https://143.244.215.182 --kubeconfig=/root/.kube/config --namespace=kubectl-4169 exec httpd -- /bin/sh -x -c /tmp/kubectl get pods --token=invalid --v=7 2>&1:
Command stdout:
I0718 19:54:06.846749     222 merged_client_builder.go:163] Using in-cluster namespace
I0718 19:54:06.846894     222 merged_client_builder.go:121] Using in-cluster configuration
I0718 19:54:06.848794     222 merged_client_builder.go:121] Using in-cluster configuration
I0718 19:54:06.852577     222 merged_client_builder.go:121] Using in-cluster configuration
I0718 19:54:06.852806     222 round_trippers.go:432] GET https://100.64.0.1:443/api/v1/namespaces/kubectl-4169/pods?limit=500
... skipping 8 lines ...
  "metadata": {},
  "status": "Failure",
  "message": "Unauthorized",
  "reason": "Unauthorized",
  "code": 401
}]
F0718 19:54:06.857963     222 helpers.go:115] error: You must be logged in to the server (Unauthorized)
goroutine 1 [running]:
k8s.io/kubernetes/vendor/k8s.io/klog/v2.stacks(0xc00000e001, 0xc0007ae000, 0x68, 0x1af)
	/workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/vendor/k8s.io/klog/v2/klog.go:1021 +0xb9
k8s.io/kubernetes/vendor/k8s.io/klog/v2.(*loggingT).output(0x3055420, 0xc000000003, 0x0, 0x0, 0xc00049ad90, 0x25f2cf0, 0xa, 0x73, 0x40e300)
	/workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/vendor/k8s.io/klog/v2/klog.go:970 +0x191
k8s.io/kubernetes/vendor/k8s.io/klog/v2.(*loggingT).printDepth(0x3055420, 0xc000000003, 0x0, 0x0, 0x0, 0x0, 0x2, 0xc000967160, 0x1, 0x1)
	/workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/vendor/k8s.io/klog/v2/klog.go:733 +0x16f
k8s.io/kubernetes/vendor/k8s.io/klog/v2.FatalDepth(...)
	/workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/vendor/k8s.io/klog/v2/klog.go:1495
k8s.io/kubernetes/vendor/k8s.io/kubectl/pkg/cmd/util.fatal(0xc00088c440, 0x3a, 0x1)
	/workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/vendor/k8s.io/kubectl/pkg/cmd/util/helpers.go:93 +0x288
k8s.io/kubernetes/vendor/k8s.io/kubectl/pkg/cmd/util.checkErr(0x207dd80, 0xc000da2630, 0x1f07e88)
	/workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/vendor/k8s.io/kubectl/pkg/cmd/util/helpers.go:177 +0x8a3
k8s.io/kubernetes/vendor/k8s.io/kubectl/pkg/cmd/util.CheckErr(...)
	/workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/vendor/k8s.io/kubectl/pkg/cmd/util/helpers.go:115
k8s.io/kubernetes/vendor/k8s.io/kubectl/pkg/cmd/get.NewCmdGet.func1(0xc00061c2c0, 0xc000526030, 0x1, 0x3)
... skipping 72 lines ...
	/workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/vendor/golang.org/x/net/http2/transport.go:2054 +0x728

stderr:
+ /tmp/kubectl get pods '--token=invalid' '--v=7'
command terminated with exit code 255

error:
exit status 255
STEP: trying to use kubectl with invalid server
Jul 18 19:54:06.881: INFO: Running '/tmp/kubectl2220469345/kubectl --server=https://143.244.215.182 --kubeconfig=/root/.kube/config --namespace=kubectl-4169 exec httpd -- /bin/sh -x -c /tmp/kubectl get pods --server=invalid --v=6 2>&1'
Jul 18 19:54:07.373: INFO: rc: 255
Jul 18 19:54:07.373: INFO: got err error running /tmp/kubectl2220469345/kubectl --server=https://143.244.215.182 --kubeconfig=/root/.kube/config --namespace=kubectl-4169 exec httpd -- /bin/sh -x -c /tmp/kubectl get pods --server=invalid --v=6 2>&1:
Command stdout:
I0718 19:54:07.332007     245 merged_client_builder.go:163] Using in-cluster namespace
I0718 19:54:07.337135     245 round_trippers.go:454] GET http://invalid/api?timeout=32s  in 4 milliseconds
I0718 19:54:07.337319     245 cached_discovery.go:121] skipped caching discovery info due to Get "http://invalid/api?timeout=32s": dial tcp: lookup invalid on 100.64.0.10:53: no such host
I0718 19:54:07.339935     245 round_trippers.go:454] GET http://invalid/api?timeout=32s  in 2 milliseconds
I0718 19:54:07.340044     245 cached_discovery.go:121] skipped caching discovery info due to Get "http://invalid/api?timeout=32s": dial tcp: lookup invalid on 100.64.0.10:53: no such host
I0718 19:54:07.340072     245 shortcut.go:89] Error loading discovery information: Get "http://invalid/api?timeout=32s": dial tcp: lookup invalid on 100.64.0.10:53: no such host
I0718 19:54:07.342306     245 round_trippers.go:454] GET http://invalid/api?timeout=32s  in 2 milliseconds
I0718 19:54:07.342345     245 cached_discovery.go:121] skipped caching discovery info due to Get "http://invalid/api?timeout=32s": dial tcp: lookup invalid on 100.64.0.10:53: no such host
I0718 19:54:07.349199     245 round_trippers.go:454] GET http://invalid/api?timeout=32s  in 6 milliseconds
I0718 19:54:07.349229     245 cached_discovery.go:121] skipped caching discovery info due to Get "http://invalid/api?timeout=32s": dial tcp: lookup invalid on 100.64.0.10:53: no such host
I0718 19:54:07.350223     245 round_trippers.go:454] GET http://invalid/api?timeout=32s  in 0 milliseconds
I0718 19:54:07.350255     245 cached_discovery.go:121] skipped caching discovery info due to Get "http://invalid/api?timeout=32s": dial tcp: lookup invalid on 100.64.0.10:53: no such host
I0718 19:54:07.350293     245 helpers.go:234] Connection error: Get http://invalid/api?timeout=32s: dial tcp: lookup invalid on 100.64.0.10:53: no such host
F0718 19:54:07.350331     245 helpers.go:115] Unable to connect to the server: dial tcp: lookup invalid on 100.64.0.10:53: no such host
goroutine 1 [running]:
k8s.io/kubernetes/vendor/k8s.io/klog/v2.stacks(0xc00013a001, 0xc000013dc0, 0x88, 0x1b8)
	/workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/vendor/k8s.io/klog/v2/klog.go:1021 +0xb9
k8s.io/kubernetes/vendor/k8s.io/klog/v2.(*loggingT).output(0x3055420, 0xc000000003, 0x0, 0x0, 0xc00086cee0, 0x25f2cf0, 0xa, 0x73, 0x40e300)
	/workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/vendor/k8s.io/klog/v2/klog.go:970 +0x191
k8s.io/kubernetes/vendor/k8s.io/klog/v2.(*loggingT).printDepth(0x3055420, 0xc000000003, 0x0, 0x0, 0x0, 0x0, 0x2, 0xc000420a90, 0x1, 0x1)
	/workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/vendor/k8s.io/klog/v2/klog.go:733 +0x16f
k8s.io/kubernetes/vendor/k8s.io/klog/v2.FatalDepth(...)
	/workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/vendor/k8s.io/klog/v2/klog.go:1495
k8s.io/kubernetes/vendor/k8s.io/kubectl/pkg/cmd/util.fatal(0xc00057e900, 0x59, 0x1)
	/workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/vendor/k8s.io/kubectl/pkg/cmd/util/helpers.go:93 +0x288
k8s.io/kubernetes/vendor/k8s.io/kubectl/pkg/cmd/util.checkErr(0x207d0e0, 0xc00056ede0, 0x1f07e88)
	/workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/vendor/k8s.io/kubectl/pkg/cmd/util/helpers.go:188 +0x935
k8s.io/kubernetes/vendor/k8s.io/kubectl/pkg/cmd/util.CheckErr(...)
	/workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/vendor/k8s.io/kubectl/pkg/cmd/util/helpers.go:115
k8s.io/kubernetes/vendor/k8s.io/kubectl/pkg/cmd/get.NewCmdGet.func1(0xc0006942c0, 0xc00042d7a0, 0x1, 0x3)
... skipping 24 lines ...
	/workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/vendor/k8s.io/kubectl/pkg/util/logs/logs.go:51 +0x96

stderr:
+ /tmp/kubectl get pods '--server=invalid' '--v=6'
command terminated with exit code 255

error:
exit status 255
STEP: trying to use kubectl with invalid namespace
Jul 18 19:54:07.373: INFO: Running '/tmp/kubectl2220469345/kubectl --server=https://143.244.215.182 --kubeconfig=/root/.kube/config --namespace=kubectl-4169 exec httpd -- /bin/sh -x -c /tmp/kubectl get pods --namespace=invalid --v=6 2>&1'
Jul 18 19:54:07.849: INFO: stderr: "+ /tmp/kubectl get pods '--namespace=invalid' '--v=6'\n"
Jul 18 19:54:07.850: INFO: stdout: "I0718 19:54:07.811459     268 merged_client_builder.go:121] Using in-cluster configuration\nI0718 19:54:07.814004     268 merged_client_builder.go:121] Using in-cluster configuration\nI0718 19:54:07.816914     268 merged_client_builder.go:121] Using in-cluster configuration\nI0718 19:54:07.826669     268 round_trippers.go:454] GET https://100.64.0.1:443/api/v1/namespaces/invalid/pods?limit=500 200 OK in 9 milliseconds\nNo resources found in invalid namespace.\n"
Jul 18 19:54:07.850: INFO: stdout: I0718 19:54:07.811459     268 merged_client_builder.go:121] Using in-cluster configuration
... skipping 104 lines ...
• [SLOW TEST:32.766 seconds]
[sig-node] Probing container
/workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/common/node/framework.go:23
  should be restarted with a /healthz http liveness probe [NodeConformance] [Conformance]
  /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:630
------------------------------
{"msg":"PASSED [sig-node] Probing container should be restarted with a /healthz http liveness probe [NodeConformance] [Conformance]","total":-1,"completed":5,"skipped":44,"failed":0}

SSSSSS
------------------------------
[BeforeEach] [sig-api-machinery] CustomResourcePublishOpenAPI [Privileged:ClusterAdmin]
  /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:185
STEP: Creating a kubernetes client
... skipping 16 lines ...
• [SLOW TEST:51.226 seconds]
[sig-api-machinery] CustomResourcePublishOpenAPI [Privileged:ClusterAdmin]
/workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/apimachinery/framework.go:23
  works for multiple CRDs of same group but different versions [Conformance]
  /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:630
------------------------------
{"msg":"PASSED [sig-api-machinery] CustomResourcePublishOpenAPI [Privileged:ClusterAdmin] works for multiple CRDs of same group but different versions [Conformance]","total":-1,"completed":5,"skipped":30,"failed":0}

SSSS
------------------------------
[BeforeEach] [sig-storage] EmptyDir wrapper volumes
  /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:185
STEP: Creating a kubernetes client
... skipping 21 lines ...
• [SLOW TEST:12.570 seconds]
[sig-storage] EmptyDir wrapper volumes
/workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/storage/utils/framework.go:23
  should not conflict [Conformance]
  /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:630
------------------------------
{"msg":"PASSED [sig-storage] EmptyDir wrapper volumes should not conflict [Conformance]","total":-1,"completed":7,"skipped":66,"failed":0}
[BeforeEach] [Testpattern: Inline-volume (ext4)] volumes
  /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/storage/framework/testsuite.go:51
Jul 18 19:54:13.174: INFO: Driver local doesn't support InlineVolume -- skipping
[AfterEach] [Testpattern: Inline-volume (ext4)] volumes
  /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:186

... skipping 180 lines ...
• [SLOW TEST:10.536 seconds]
[sig-node] Pods
/workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/common/node/framework.go:23
  should support remote command execution over websockets [NodeConformance] [Conformance]
  /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:630
------------------------------
{"msg":"PASSED [sig-node] Pods should support remote command execution over websockets [NodeConformance] [Conformance]","total":-1,"completed":8,"skipped":72,"failed":0}

SS
------------------------------
[BeforeEach] [sig-storage] Downward API volume
  /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:185
STEP: Creating a kubernetes client
... skipping 20 lines ...
• [SLOW TEST:11.063 seconds]
[sig-storage] Downward API volume
/workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/common/storage/framework.go:23
  should update annotations on modification [NodeConformance] [Conformance]
  /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:630
------------------------------
{"msg":"PASSED [sig-storage] Downward API volume should update annotations on modification [NodeConformance] [Conformance]","total":-1,"completed":9,"skipped":20,"failed":0}
[BeforeEach] [Testpattern: Inline-volume (ext4)] volumes
  /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/storage/framework/testsuite.go:51
Jul 18 19:54:15.868: INFO: Driver hostPathSymlink doesn't support ext4 -- skipping
[AfterEach] [Testpattern: Inline-volume (ext4)] volumes
  /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:186

... skipping 56 lines ...
• [SLOW TEST:10.503 seconds]
[sig-storage] ConfigMap
/workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/common/storage/framework.go:23
  binary data should be reflected in volume [NodeConformance] [Conformance]
  /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:630
------------------------------
{"msg":"PASSED [sig-storage] ConfigMap binary data should be reflected in volume [NodeConformance] [Conformance]","total":-1,"completed":8,"skipped":30,"failed":0}

SSSSSS
------------------------------
[BeforeEach] [Testpattern: Dynamic PV (default fs)] subPath
  /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/storage/framework/testsuite.go:51
[BeforeEach] [Testpattern: Dynamic PV (default fs)] subPath
... skipping 51 lines ...
Jul 18 19:52:55.413: INFO: Waiting up to timeout=5m0s for PersistentVolumeClaims [csi-hostpathl72wj] to have phase Bound
Jul 18 19:52:55.443: INFO: PersistentVolumeClaim csi-hostpathl72wj found but phase is Pending instead of Bound.
Jul 18 19:52:57.477: INFO: PersistentVolumeClaim csi-hostpathl72wj found but phase is Pending instead of Bound.
Jul 18 19:52:59.508: INFO: PersistentVolumeClaim csi-hostpathl72wj found and phase=Bound (4.095602227s)
STEP: Creating pod pod-subpath-test-dynamicpv-fvk2
STEP: Creating a pod to test atomic-volume-subpath
Jul 18 19:52:59.609: INFO: Waiting up to 5m0s for pod "pod-subpath-test-dynamicpv-fvk2" in namespace "provisioning-8290" to be "Succeeded or Failed"
Jul 18 19:52:59.640: INFO: Pod "pod-subpath-test-dynamicpv-fvk2": Phase="Pending", Reason="", readiness=false. Elapsed: 30.255176ms
Jul 18 19:53:01.672: INFO: Pod "pod-subpath-test-dynamicpv-fvk2": Phase="Pending", Reason="", readiness=false. Elapsed: 2.062344376s
Jul 18 19:53:03.706: INFO: Pod "pod-subpath-test-dynamicpv-fvk2": Phase="Pending", Reason="", readiness=false. Elapsed: 4.096090424s
Jul 18 19:53:05.745: INFO: Pod "pod-subpath-test-dynamicpv-fvk2": Phase="Pending", Reason="", readiness=false. Elapsed: 6.135337942s
Jul 18 19:53:07.777: INFO: Pod "pod-subpath-test-dynamicpv-fvk2": Phase="Pending", Reason="", readiness=false. Elapsed: 8.167184751s
Jul 18 19:53:09.807: INFO: Pod "pod-subpath-test-dynamicpv-fvk2": Phase="Pending", Reason="", readiness=false. Elapsed: 10.197869642s
... skipping 12 lines ...
Jul 18 19:53:36.232: INFO: Pod "pod-subpath-test-dynamicpv-fvk2": Phase="Running", Reason="", readiness=true. Elapsed: 36.622598133s
Jul 18 19:53:38.265: INFO: Pod "pod-subpath-test-dynamicpv-fvk2": Phase="Running", Reason="", readiness=true. Elapsed: 38.655901053s
Jul 18 19:53:40.298: INFO: Pod "pod-subpath-test-dynamicpv-fvk2": Phase="Running", Reason="", readiness=true. Elapsed: 40.688966051s
Jul 18 19:53:42.331: INFO: Pod "pod-subpath-test-dynamicpv-fvk2": Phase="Running", Reason="", readiness=true. Elapsed: 42.721649016s
Jul 18 19:53:44.362: INFO: Pod "pod-subpath-test-dynamicpv-fvk2": Phase="Succeeded", Reason="", readiness=false. Elapsed: 44.75299948s
STEP: Saw pod success
Jul 18 19:53:44.363: INFO: Pod "pod-subpath-test-dynamicpv-fvk2" satisfied condition "Succeeded or Failed"
Jul 18 19:53:44.396: INFO: Trying to get logs from node 10.116.0.7 pod pod-subpath-test-dynamicpv-fvk2 container test-container-subpath-dynamicpv-fvk2: <nil>
STEP: delete the pod
Jul 18 19:53:44.478: INFO: Waiting for pod pod-subpath-test-dynamicpv-fvk2 to disappear
Jul 18 19:53:44.509: INFO: Pod pod-subpath-test-dynamicpv-fvk2 no longer exists
STEP: Deleting pod pod-subpath-test-dynamicpv-fvk2
Jul 18 19:53:44.509: INFO: Deleting pod "pod-subpath-test-dynamicpv-fvk2" in namespace "provisioning-8290"
... skipping 54 lines ...
  /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/storage/csi_volumes.go:40
    [Testpattern: Dynamic PV (default fs)] subPath
    /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/storage/framework/testsuite.go:50
      should support file as subpath [LinuxOnly]
      /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/storage/testsuites/subpath.go:230
------------------------------
{"msg":"PASSED [sig-storage] CSI Volumes [Driver: csi-hostpath] [Testpattern: Dynamic PV (default fs)] subPath should support file as subpath [LinuxOnly]","total":-1,"completed":9,"skipped":73,"failed":0}

S
------------------------------
{"msg":"PASSED [sig-cli] Kubectl client Simple pod should handle in-cluster config","total":-1,"completed":7,"skipped":55,"failed":0}
[BeforeEach] [sig-node] Variable Expansion
  /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:185
STEP: Creating a kubernetes client
Jul 18 19:54:09.039: INFO: >>> kubeConfig: /root/.kube/config
STEP: Building a namespace api object, basename var-expansion
STEP: Waiting for a default service account to be provisioned in namespace
[It] should allow composing env vars into new env vars [NodeConformance] [Conformance]
  /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:630
STEP: Creating a pod to test env composition
Jul 18 19:54:09.242: INFO: Waiting up to 5m0s for pod "var-expansion-2fab8eaa-ab91-49ff-aa73-508ebfecd442" in namespace "var-expansion-5857" to be "Succeeded or Failed"
Jul 18 19:54:09.275: INFO: Pod "var-expansion-2fab8eaa-ab91-49ff-aa73-508ebfecd442": Phase="Pending", Reason="", readiness=false. Elapsed: 32.143038ms
Jul 18 19:54:11.312: INFO: Pod "var-expansion-2fab8eaa-ab91-49ff-aa73-508ebfecd442": Phase="Pending", Reason="", readiness=false. Elapsed: 2.069521916s
Jul 18 19:54:13.346: INFO: Pod "var-expansion-2fab8eaa-ab91-49ff-aa73-508ebfecd442": Phase="Pending", Reason="", readiness=false. Elapsed: 4.103624528s
Jul 18 19:54:15.379: INFO: Pod "var-expansion-2fab8eaa-ab91-49ff-aa73-508ebfecd442": Phase="Pending", Reason="", readiness=false. Elapsed: 6.13630061s
Jul 18 19:54:17.411: INFO: Pod "var-expansion-2fab8eaa-ab91-49ff-aa73-508ebfecd442": Phase="Succeeded", Reason="", readiness=false. Elapsed: 8.168108823s
STEP: Saw pod success
Jul 18 19:54:17.411: INFO: Pod "var-expansion-2fab8eaa-ab91-49ff-aa73-508ebfecd442" satisfied condition "Succeeded or Failed"
Jul 18 19:54:17.444: INFO: Trying to get logs from node 10.116.0.5 pod var-expansion-2fab8eaa-ab91-49ff-aa73-508ebfecd442 container dapi-container: <nil>
STEP: delete the pod
Jul 18 19:54:17.538: INFO: Waiting for pod var-expansion-2fab8eaa-ab91-49ff-aa73-508ebfecd442 to disappear
Jul 18 19:54:17.569: INFO: Pod var-expansion-2fab8eaa-ab91-49ff-aa73-508ebfecd442 no longer exists
[AfterEach] [sig-node] Variable Expansion
  /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:186
... skipping 4 lines ...
• [SLOW TEST:8.635 seconds]
[sig-node] Variable Expansion
/workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/common/node/framework.go:23
  should allow composing env vars into new env vars [NodeConformance] [Conformance]
  /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:630
------------------------------
{"msg":"PASSED [sig-node] Variable Expansion should allow composing env vars into new env vars [NodeConformance] [Conformance]","total":-1,"completed":8,"skipped":55,"failed":0}

SSSSSSS
------------------------------
[BeforeEach] [Testpattern: Pre-provisioned PV (default fs)] subPath
  /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/storage/framework/testsuite.go:51
Jul 18 19:54:17.711: INFO: Driver hostPathSymlink doesn't support PreprovisionedPV -- skipping
... skipping 40 lines ...
• [SLOW TEST:9.067 seconds]
[sig-storage] Projected downwardAPI
/workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/common/storage/framework.go:23
  should update labels on modification [NodeConformance] [Conformance]
  /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:630
------------------------------
{"msg":"PASSED [sig-storage] Projected downwardAPI should update labels on modification [NodeConformance] [Conformance]","total":-1,"completed":6,"skipped":34,"failed":0}

SSSS
------------------------------
[BeforeEach] [Testpattern: Inline-volume (ext4)] volumes
  /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/storage/framework/testsuite.go:51
Jul 18 19:54:21.047: INFO: Only supported for providers [vsphere] (not skeleton)
... skipping 39 lines ...
• [SLOW TEST:43.271 seconds]
[sig-api-machinery] Garbage collector
/workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/apimachinery/framework.go:23
  should support orphan deletion of custom resources
  /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/apimachinery/garbage_collector.go:1055
------------------------------
{"msg":"PASSED [sig-api-machinery] Garbage collector should support orphan deletion of custom resources","total":-1,"completed":11,"skipped":56,"failed":0}
[BeforeEach] [Testpattern: Dynamic PV (default fs)] subPath
  /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/storage/framework/testsuite.go:51
Jul 18 19:54:27.160: INFO: Only supported for providers [gce gke] (not skeleton)
[AfterEach] [Testpattern: Dynamic PV (default fs)] subPath
  /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:186

... skipping 113 lines ...
Jul 18 19:53:27.674: INFO: PersistentVolumeClaim pvc-mz5wp found and phase=Bound (2.064528622s)
STEP: Deleting the previously created pod
Jul 18 19:53:42.840: INFO: Deleting pod "pvc-volume-tester-z462c" in namespace "csi-mock-volumes-523"
Jul 18 19:53:42.877: INFO: Wait up to 5m0s for pod "pvc-volume-tester-z462c" to be fully deleted
STEP: Checking CSI driver logs
Jul 18 19:53:50.975: INFO: Found volume attribute csi.storage.k8s.io/serviceAccount.tokens: {"":{"token":"eyJhbGciOiJSUzI1NiIsImtpZCI6Inc1b19LRUtfVXdKYXpSMWdDU1MyUU93Q1dzN3dLT2ktT2tMbTV0dThyMFUifQ.eyJhdWQiOlsia3ViZXJuZXRlcy5zdmMuZGVmYXVsdCJdLCJleHAiOjE2MjY2Mzg2MTUsImlhdCI6MTYyNjYzODAxNSwiaXNzIjoiaHR0cHM6Ly9rdWJlcm5ldGVzLmRlZmF1bHQiLCJrdWJlcm5ldGVzLmlvIjp7Im5hbWVzcGFjZSI6ImNzaS1tb2NrLXZvbHVtZXMtNTIzIiwicG9kIjp7Im5hbWUiOiJwdmMtdm9sdW1lLXRlc3Rlci16NDYyYyIsInVpZCI6ImQ4MWZkOTY4LWU0YTItNGM2Ny05NGU1LWZhMGZlZWI4YzlmMSJ9LCJzZXJ2aWNlYWNjb3VudCI6eyJuYW1lIjoiZGVmYXVsdCIsInVpZCI6Ijc3NzNiM2UwLTU4NmItNGIxMy04NWM5LWY5OGY2MTM5ZjUzYyJ9fSwibmJmIjoxNjI2NjM4MDE1LCJzdWIiOiJzeXN0ZW06c2VydmljZWFjY291bnQ6Y3NpLW1vY2stdm9sdW1lcy01MjM6ZGVmYXVsdCJ9.wG49McYk7OEMmsXNtpYQDepqe9QS43k6n7Gb2biIPedOyVlDjOwuQ1ywlLcI-eVWNDwW_uZ90YRw7b2-Tkkxx7ZSDgyLWLkGDxByFds5vdlNoy9oKBbwqrzU70ealxr-mA7eEbNWUMGOPMv71_t43M5Kx5-PDcp384h6GVJHZn5_yGZNpuAWZSoIGHiF4WV7O9V0dvePlyWJZ598ahTC4V8bQ3wg5PUp_qkmh_N6YdfZIEd65LdWGvgvXLqX_KvJxtUSG0QPOssnEJhGYwDv2ByrovcKmoW_bNbehYFL5U2UN2aJP2ohz8NI4v_AmaJyF7Hdq6yAMcQMmevrgOW-8A","expirationTimestamp":"2021-07-18T20:03:35Z"}}
Jul 18 19:53:50.975: INFO: Found NodeUnpublishVolume: {json: {"Method":"/csi.v1.Node/NodeUnpublishVolume","Request":{"volume_id":"4","target_path":"/var/lib/kubelet/pods/d81fd968-e4a2-4c67-94e5-fa0feeb8c9f1/volumes/kubernetes.io~csi/pvc-dcca0ff4-ffed-42df-866c-0b29aa859076/mount"},"Response":{},"Error":"","FullError":null} Method:NodeUnpublishVolume Request:{VolumeContext:map[]} FullError:{Code:OK Message:} Error:}
STEP: Deleting pod pvc-volume-tester-z462c
Jul 18 19:53:50.975: INFO: Deleting pod "pvc-volume-tester-z462c" in namespace "csi-mock-volumes-523"
STEP: Deleting claim pvc-mz5wp
Jul 18 19:53:51.070: INFO: Waiting up to 2m0s for PersistentVolume pvc-dcca0ff4-ffed-42df-866c-0b29aa859076 to get deleted
Jul 18 19:53:51.100: INFO: PersistentVolume pvc-dcca0ff4-ffed-42df-866c-0b29aa859076 found and phase=Released (29.682837ms)
Jul 18 19:53:53.131: INFO: PersistentVolume pvc-dcca0ff4-ffed-42df-866c-0b29aa859076 was removed
... skipping 45 lines ...
/workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/storage/utils/framework.go:23
  CSIServiceAccountToken
  /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/storage/csi_mock_volume.go:1374
    token should be plumbed down when csiServiceAccountTokenEnabled=true
    /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/storage/csi_mock_volume.go:1402
------------------------------
{"msg":"PASSED [sig-storage] CSI mock volume CSIServiceAccountToken token should be plumbed down when csiServiceAccountTokenEnabled=true","total":-1,"completed":6,"skipped":20,"failed":0}

SS
------------------------------
[BeforeEach] [Testpattern: Pre-provisioned PV (default fs)] subPath
  /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/storage/framework/testsuite.go:51
Jul 18 19:54:28.587: INFO: Driver hostPathSymlink doesn't support PreprovisionedPV -- skipping
... skipping 74 lines ...
[It] should allow exec of files on the volume
  /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/storage/testsuites/volumes.go:196
Jul 18 19:54:17.588: INFO: In-tree plugin kubernetes.io/empty-dir is not migrated, not validating any metrics
Jul 18 19:54:17.588: INFO: Creating resource for inline volume
STEP: Creating pod exec-volume-test-inlinevolume-frcx
STEP: Creating a pod to test exec-volume-test
Jul 18 19:54:17.630: INFO: Waiting up to 5m0s for pod "exec-volume-test-inlinevolume-frcx" in namespace "volume-1445" to be "Succeeded or Failed"
Jul 18 19:54:17.661: INFO: Pod "exec-volume-test-inlinevolume-frcx": Phase="Pending", Reason="", readiness=false. Elapsed: 31.397537ms
Jul 18 19:54:19.693: INFO: Pod "exec-volume-test-inlinevolume-frcx": Phase="Pending", Reason="", readiness=false. Elapsed: 2.062960037s
Jul 18 19:54:21.725: INFO: Pod "exec-volume-test-inlinevolume-frcx": Phase="Pending", Reason="", readiness=false. Elapsed: 4.095413489s
Jul 18 19:54:23.757: INFO: Pod "exec-volume-test-inlinevolume-frcx": Phase="Pending", Reason="", readiness=false. Elapsed: 6.12706745s
Jul 18 19:54:25.788: INFO: Pod "exec-volume-test-inlinevolume-frcx": Phase="Pending", Reason="", readiness=false. Elapsed: 8.158331822s
Jul 18 19:54:27.820: INFO: Pod "exec-volume-test-inlinevolume-frcx": Phase="Pending", Reason="", readiness=false. Elapsed: 10.189659986s
Jul 18 19:54:29.850: INFO: Pod "exec-volume-test-inlinevolume-frcx": Phase="Succeeded", Reason="", readiness=false. Elapsed: 12.220409841s
STEP: Saw pod success
Jul 18 19:54:29.850: INFO: Pod "exec-volume-test-inlinevolume-frcx" satisfied condition "Succeeded or Failed"
Jul 18 19:54:29.889: INFO: Trying to get logs from node 10.116.0.5 pod exec-volume-test-inlinevolume-frcx container exec-container-inlinevolume-frcx: <nil>
STEP: delete the pod
Jul 18 19:54:29.964: INFO: Waiting for pod exec-volume-test-inlinevolume-frcx to disappear
Jul 18 19:54:29.996: INFO: Pod exec-volume-test-inlinevolume-frcx no longer exists
STEP: Deleting pod exec-volume-test-inlinevolume-frcx
Jul 18 19:54:29.996: INFO: Deleting pod "exec-volume-test-inlinevolume-frcx" in namespace "volume-1445"
... skipping 10 lines ...
  /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/storage/in_tree_volumes.go:58
    [Testpattern: Inline-volume (default fs)] volumes
    /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/storage/framework/testsuite.go:50
      should allow exec of files on the volume
      /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/storage/testsuites/volumes.go:196
------------------------------
{"msg":"PASSED [sig-storage] In-tree Volumes [Driver: emptydir] [Testpattern: Inline-volume (default fs)] volumes should allow exec of files on the volume","total":-1,"completed":9,"skipped":36,"failed":0}

SSS
------------------------------
[BeforeEach] [Testpattern: Dynamic PV (default fs)] fsgroupchangepolicy
  /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/storage/framework/testsuite.go:51
Jul 18 19:54:30.141: INFO: Driver local doesn't support DynamicPV -- skipping
... skipping 156 lines ...
Jul 18 19:54:05.296: INFO: >>> kubeConfig: /root/.kube/config
Jul 18 19:54:05.549: INFO: Exec stderr: ""
Jul 18 19:54:23.646: INFO: ExecWithOptions {Command:[nsenter --mount=/rootfs/proc/1/ns/mnt -- sh -c mkdir "/var/lib/kubelet/mount-propagation-6966"/host; mount -t tmpfs e2e-mount-propagation-host "/var/lib/kubelet/mount-propagation-6966"/host; echo host > "/var/lib/kubelet/mount-propagation-6966"/host/file] Namespace:mount-propagation-6966 PodName:hostexec-10.116.0.7-psl9z ContainerName:agnhost-container Stdin:<nil> CaptureStdout:true CaptureStderr:true PreserveWhitespace:true Quiet:false}
Jul 18 19:54:23.646: INFO: >>> kubeConfig: /root/.kube/config
Jul 18 19:54:23.929: INFO: ExecWithOptions {Command:[/bin/sh -c cat /mnt/test/master/file] Namespace:mount-propagation-6966 PodName:private ContainerName:cntr Stdin:<nil> CaptureStdout:true CaptureStderr:true PreserveWhitespace:false Quiet:false}
Jul 18 19:54:23.929: INFO: >>> kubeConfig: /root/.kube/config
Jul 18 19:54:24.183: INFO: pod private mount master: stdout: "", stderr: "cat: can't open '/mnt/test/master/file': No such file or directory" error: command terminated with exit code 1
Jul 18 19:54:24.214: INFO: ExecWithOptions {Command:[/bin/sh -c cat /mnt/test/slave/file] Namespace:mount-propagation-6966 PodName:private ContainerName:cntr Stdin:<nil> CaptureStdout:true CaptureStderr:true PreserveWhitespace:false Quiet:false}
Jul 18 19:54:24.214: INFO: >>> kubeConfig: /root/.kube/config
Jul 18 19:54:24.458: INFO: pod private mount slave: stdout: "", stderr: "cat: can't open '/mnt/test/slave/file': No such file or directory" error: command terminated with exit code 1
Jul 18 19:54:24.488: INFO: ExecWithOptions {Command:[/bin/sh -c cat /mnt/test/private/file] Namespace:mount-propagation-6966 PodName:private ContainerName:cntr Stdin:<nil> CaptureStdout:true CaptureStderr:true PreserveWhitespace:false Quiet:false}
Jul 18 19:54:24.488: INFO: >>> kubeConfig: /root/.kube/config
Jul 18 19:54:24.733: INFO: pod private mount private: stdout: "private", stderr: "" error: <nil>
Jul 18 19:54:24.764: INFO: ExecWithOptions {Command:[/bin/sh -c cat /mnt/test/default/file] Namespace:mount-propagation-6966 PodName:private ContainerName:cntr Stdin:<nil> CaptureStdout:true CaptureStderr:true PreserveWhitespace:false Quiet:false}
Jul 18 19:54:24.764: INFO: >>> kubeConfig: /root/.kube/config
Jul 18 19:54:25.021: INFO: pod private mount default: stdout: "", stderr: "cat: can't open '/mnt/test/default/file': No such file or directory" error: command terminated with exit code 1
Jul 18 19:54:25.052: INFO: ExecWithOptions {Command:[/bin/sh -c cat /mnt/test/host/file] Namespace:mount-propagation-6966 PodName:private ContainerName:cntr Stdin:<nil> CaptureStdout:true CaptureStderr:true PreserveWhitespace:false Quiet:false}
Jul 18 19:54:25.052: INFO: >>> kubeConfig: /root/.kube/config
Jul 18 19:54:25.307: INFO: pod private mount host: stdout: "", stderr: "cat: can't open '/mnt/test/host/file': No such file or directory" error: command terminated with exit code 1
Jul 18 19:54:25.337: INFO: ExecWithOptions {Command:[/bin/sh -c cat /mnt/test/master/file] Namespace:mount-propagation-6966 PodName:default ContainerName:cntr Stdin:<nil> CaptureStdout:true CaptureStderr:true PreserveWhitespace:false Quiet:false}
Jul 18 19:54:25.337: INFO: >>> kubeConfig: /root/.kube/config
Jul 18 19:54:25.562: INFO: pod default mount master: stdout: "", stderr: "cat: can't open '/mnt/test/master/file': No such file or directory" error: command terminated with exit code 1
Jul 18 19:54:25.593: INFO: ExecWithOptions {Command:[/bin/sh -c cat /mnt/test/slave/file] Namespace:mount-propagation-6966 PodName:default ContainerName:cntr Stdin:<nil> CaptureStdout:true CaptureStderr:true PreserveWhitespace:false Quiet:false}
Jul 18 19:54:25.593: INFO: >>> kubeConfig: /root/.kube/config
Jul 18 19:54:25.816: INFO: pod default mount slave: stdout: "", stderr: "cat: can't open '/mnt/test/slave/file': No such file or directory" error: command terminated with exit code 1
Jul 18 19:54:25.847: INFO: ExecWithOptions {Command:[/bin/sh -c cat /mnt/test/private/file] Namespace:mount-propagation-6966 PodName:default ContainerName:cntr Stdin:<nil> CaptureStdout:true CaptureStderr:true PreserveWhitespace:false Quiet:false}
Jul 18 19:54:25.847: INFO: >>> kubeConfig: /root/.kube/config
Jul 18 19:54:26.100: INFO: pod default mount private: stdout: "", stderr: "cat: can't open '/mnt/test/private/file': No such file or directory" error: command terminated with exit code 1
Jul 18 19:54:26.131: INFO: ExecWithOptions {Command:[/bin/sh -c cat /mnt/test/default/file] Namespace:mount-propagation-6966 PodName:default ContainerName:cntr Stdin:<nil> CaptureStdout:true CaptureStderr:true PreserveWhitespace:false Quiet:false}
Jul 18 19:54:26.131: INFO: >>> kubeConfig: /root/.kube/config
Jul 18 19:54:26.388: INFO: pod default mount default: stdout: "default", stderr: "" error: <nil>
Jul 18 19:54:26.419: INFO: ExecWithOptions {Command:[/bin/sh -c cat /mnt/test/host/file] Namespace:mount-propagation-6966 PodName:default ContainerName:cntr Stdin:<nil> CaptureStdout:true CaptureStderr:true PreserveWhitespace:false Quiet:false}
Jul 18 19:54:26.419: INFO: >>> kubeConfig: /root/.kube/config
Jul 18 19:54:26.671: INFO: pod default mount host: stdout: "", stderr: "cat: can't open '/mnt/test/host/file': No such file or directory" error: command terminated with exit code 1
Jul 18 19:54:26.702: INFO: ExecWithOptions {Command:[/bin/sh -c cat /mnt/test/master/file] Namespace:mount-propagation-6966 PodName:master ContainerName:cntr Stdin:<nil> CaptureStdout:true CaptureStderr:true PreserveWhitespace:false Quiet:false}
Jul 18 19:54:26.702: INFO: >>> kubeConfig: /root/.kube/config
Jul 18 19:54:26.966: INFO: pod master mount master: stdout: "master", stderr: "" error: <nil>
Jul 18 19:54:26.997: INFO: ExecWithOptions {Command:[/bin/sh -c cat /mnt/test/slave/file] Namespace:mount-propagation-6966 PodName:master ContainerName:cntr Stdin:<nil> CaptureStdout:true CaptureStderr:true PreserveWhitespace:false Quiet:false}
Jul 18 19:54:26.997: INFO: >>> kubeConfig: /root/.kube/config
Jul 18 19:54:27.241: INFO: pod master mount slave: stdout: "", stderr: "cat: can't open '/mnt/test/slave/file': No such file or directory" error: command terminated with exit code 1
Jul 18 19:54:27.272: INFO: ExecWithOptions {Command:[/bin/sh -c cat /mnt/test/private/file] Namespace:mount-propagation-6966 PodName:master ContainerName:cntr Stdin:<nil> CaptureStdout:true CaptureStderr:true PreserveWhitespace:false Quiet:false}
Jul 18 19:54:27.272: INFO: >>> kubeConfig: /root/.kube/config
Jul 18 19:54:27.496: INFO: pod master mount private: stdout: "", stderr: "cat: can't open '/mnt/test/private/file': No such file or directory" error: command terminated with exit code 1
Jul 18 19:54:27.538: INFO: ExecWithOptions {Command:[/bin/sh -c cat /mnt/test/default/file] Namespace:mount-propagation-6966 PodName:master ContainerName:cntr Stdin:<nil> CaptureStdout:true CaptureStderr:true PreserveWhitespace:false Quiet:false}
Jul 18 19:54:27.538: INFO: >>> kubeConfig: /root/.kube/config
Jul 18 19:54:27.776: INFO: pod master mount default: stdout: "", stderr: "cat: can't open '/mnt/test/default/file': No such file or directory" error: command terminated with exit code 1
Jul 18 19:54:27.807: INFO: ExecWithOptions {Command:[/bin/sh -c cat /mnt/test/host/file] Namespace:mount-propagation-6966 PodName:master ContainerName:cntr Stdin:<nil> CaptureStdout:true CaptureStderr:true PreserveWhitespace:false Quiet:false}
Jul 18 19:54:27.807: INFO: >>> kubeConfig: /root/.kube/config
Jul 18 19:54:28.032: INFO: pod master mount host: stdout: "host", stderr: "" error: <nil>
Jul 18 19:54:28.062: INFO: ExecWithOptions {Command:[/bin/sh -c cat /mnt/test/master/file] Namespace:mount-propagation-6966 PodName:slave ContainerName:cntr Stdin:<nil> CaptureStdout:true CaptureStderr:true PreserveWhitespace:false Quiet:false}
Jul 18 19:54:28.062: INFO: >>> kubeConfig: /root/.kube/config
Jul 18 19:54:28.319: INFO: pod slave mount master: stdout: "master", stderr: "" error: <nil>
Jul 18 19:54:28.350: INFO: ExecWithOptions {Command:[/bin/sh -c cat /mnt/test/slave/file] Namespace:mount-propagation-6966 PodName:slave ContainerName:cntr Stdin:<nil> CaptureStdout:true CaptureStderr:true PreserveWhitespace:false Quiet:false}
Jul 18 19:54:28.350: INFO: >>> kubeConfig: /root/.kube/config
Jul 18 19:54:28.705: INFO: pod slave mount slave: stdout: "slave", stderr: "" error: <nil>
Jul 18 19:54:28.736: INFO: ExecWithOptions {Command:[/bin/sh -c cat /mnt/test/private/file] Namespace:mount-propagation-6966 PodName:slave ContainerName:cntr Stdin:<nil> CaptureStdout:true CaptureStderr:true PreserveWhitespace:false Quiet:false}
Jul 18 19:54:28.736: INFO: >>> kubeConfig: /root/.kube/config
Jul 18 19:54:28.983: INFO: pod slave mount private: stdout: "", stderr: "cat: can't open '/mnt/test/private/file': No such file or directory" error: command terminated with exit code 1
Jul 18 19:54:29.014: INFO: ExecWithOptions {Command:[/bin/sh -c cat /mnt/test/default/file] Namespace:mount-propagation-6966 PodName:slave ContainerName:cntr Stdin:<nil> CaptureStdout:true CaptureStderr:true PreserveWhitespace:false Quiet:false}
Jul 18 19:54:29.015: INFO: >>> kubeConfig: /root/.kube/config
Jul 18 19:54:29.265: INFO: pod slave mount default: stdout: "", stderr: "cat: can't open '/mnt/test/default/file': No such file or directory" error: command terminated with exit code 1
Jul 18 19:54:29.296: INFO: ExecWithOptions {Command:[/bin/sh -c cat /mnt/test/host/file] Namespace:mount-propagation-6966 PodName:slave ContainerName:cntr Stdin:<nil> CaptureStdout:true CaptureStderr:true PreserveWhitespace:false Quiet:false}
Jul 18 19:54:29.296: INFO: >>> kubeConfig: /root/.kube/config
Jul 18 19:54:29.565: INFO: pod slave mount host: stdout: "host", stderr: "" error: <nil>
Jul 18 19:54:29.565: INFO: ExecWithOptions {Command:[nsenter --mount=/rootfs/proc/1/ns/mnt -- sh -c test `cat "/var/lib/kubelet/mount-propagation-6966"/master/file` = master] Namespace:mount-propagation-6966 PodName:hostexec-10.116.0.7-psl9z ContainerName:agnhost-container Stdin:<nil> CaptureStdout:true CaptureStderr:true PreserveWhitespace:true Quiet:false}
Jul 18 19:54:29.566: INFO: >>> kubeConfig: /root/.kube/config
Jul 18 19:54:29.816: INFO: ExecWithOptions {Command:[nsenter --mount=/rootfs/proc/1/ns/mnt -- sh -c test ! -e "/var/lib/kubelet/mount-propagation-6966"/slave/file] Namespace:mount-propagation-6966 PodName:hostexec-10.116.0.7-psl9z ContainerName:agnhost-container Stdin:<nil> CaptureStdout:true CaptureStderr:true PreserveWhitespace:true Quiet:false}
Jul 18 19:54:29.816: INFO: >>> kubeConfig: /root/.kube/config
Jul 18 19:54:30.069: INFO: ExecWithOptions {Command:[nsenter --mount=/rootfs/proc/1/ns/mnt -- sh -c umount "/var/lib/kubelet/mount-propagation-6966"/host] Namespace:mount-propagation-6966 PodName:hostexec-10.116.0.7-psl9z ContainerName:agnhost-container Stdin:<nil> CaptureStdout:true CaptureStderr:true PreserveWhitespace:true Quiet:false}
Jul 18 19:54:30.069: INFO: >>> kubeConfig: /root/.kube/config
... skipping 21 lines ...
• [SLOW TEST:72.738 seconds]
[sig-node] Mount propagation
/workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/node/framework.go:23
  should propagate mounts to the host
  /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/node/mount_propagation.go:82
------------------------------
{"msg":"PASSED [sig-node] Mount propagation should propagate mounts to the host","total":-1,"completed":5,"skipped":18,"failed":0}

S
------------------------------
[BeforeEach] [Testpattern: Inline-volume (default fs)] subPath
  /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/storage/framework/testsuite.go:51
Jul 18 19:54:31.845: INFO: Only supported for node OS distro [gci ubuntu custom] (not debian)
... skipping 59 lines ...
• [SLOW TEST:16.499 seconds]
[sig-apps] DisruptionController
/workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/apps/framework.go:23
  evictions: too few pods, absolute => should not allow an eviction
  /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/apps/disruption.go:267
------------------------------
{"msg":"PASSED [sig-apps] DisruptionController evictions: too few pods, absolute =\u003e should not allow an eviction","total":-1,"completed":10,"skipped":23,"failed":0}

SSSSSS
------------------------------
[BeforeEach] [Testpattern: Dynamic PV (default fs)] subPath
  /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/storage/framework/testsuite.go:51
Jul 18 19:54:32.422: INFO: Only supported for providers [aws] (not skeleton)
... skipping 49 lines ...
[sig-storage] In-tree Volumes
/workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/storage/utils/framework.go:23
  [Driver: local][LocalVolumeType: dir-link]
  /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/storage/in_tree_volumes.go:58
    [Testpattern: Dynamic PV (delayed binding)] topology
    /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/storage/framework/testsuite.go:50
      should fail to schedule a pod which has topologies that conflict with AllowedTopologies [BeforeEach]
      /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/storage/testsuites/topology.go:192

      Driver local doesn't support DynamicPV -- skipping

      /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/storage/framework/testsuite.go:116
------------------------------
... skipping 43 lines ...
• [SLOW TEST:19.481 seconds]
[sig-apps] ReplicationController
/workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/apps/framework.go:23
  should serve a basic image on each replica with a public image  [Conformance]
  /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:630
------------------------------
{"msg":"PASSED [sig-apps] ReplicationController should serve a basic image on each replica with a public image  [Conformance]","total":-1,"completed":8,"skipped":85,"failed":0}

S
------------------------------
[BeforeEach] [Testpattern: Pre-provisioned PV (ext3)] volumes
  /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/storage/framework/testsuite.go:51
Jul 18 19:54:32.778: INFO: Only supported for providers [openstack] (not skeleton)
... skipping 115 lines ...
  /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/storage/persistent_volumes-local.go:192
    Two pods mounting a local volume at the same time
    /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/storage/persistent_volumes-local.go:248
      should be able to write from pod1 and read from pod2
      /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/storage/persistent_volumes-local.go:249
------------------------------
{"msg":"PASSED [sig-storage] PersistentVolumes-local  [Volume type: dir-link] Two pods mounting a local volume at the same time should be able to write from pod1 and read from pod2","total":-1,"completed":6,"skipped":50,"failed":0}

SSSSS
------------------------------
[BeforeEach] [Testpattern: Pre-provisioned PV (default fs)] subPath
  /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/storage/framework/testsuite.go:51
Jul 18 19:54:33.082: INFO: Only supported for providers [azure] (not skeleton)
... skipping 78 lines ...
  /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:186
Jul 18 19:54:33.463: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready
STEP: Destroying namespace "kubectl-5070" for this suite.

•
------------------------------
{"msg":"PASSED [sig-cli] Kubectl client Proxy server should support proxy with --port 0  [Conformance]","total":-1,"completed":7,"skipped":70,"failed":0}

S
------------------------------
[BeforeEach] [Testpattern: Dynamic PV (ext4)] volumes
  /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/storage/framework/testsuite.go:51
Jul 18 19:54:33.546: INFO: Only supported for providers [azure] (not skeleton)
... skipping 37 lines ...
• [SLOW TEST:17.065 seconds]
[sig-api-machinery] ServerSideApply
/workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/apimachinery/framework.go:23
  should work for CRDs
  /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/apimachinery/apply.go:569
------------------------------
{"msg":"PASSED [sig-api-machinery] ServerSideApply should work for CRDs","total":-1,"completed":7,"skipped":42,"failed":0}
[BeforeEach] [Testpattern: Dynamic PV (default fs)] fsgroupchangepolicy
  /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/storage/framework/testsuite.go:51
Jul 18 19:54:38.142: INFO: Only supported for providers [vsphere] (not skeleton)
[AfterEach] [Testpattern: Dynamic PV (default fs)] fsgroupchangepolicy
  /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:186

... skipping 48 lines ...
• [SLOW TEST:10.678 seconds]
[sig-api-machinery] Watchers
/workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/apimachinery/framework.go:23
  should observe an object deletion if it stops meeting the requirements of the selector [Conformance]
  /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:630
------------------------------
{"msg":"PASSED [sig-api-machinery] Watchers should observe an object deletion if it stops meeting the requirements of the selector [Conformance]","total":-1,"completed":7,"skipped":41,"failed":0}

S
------------------------------
[BeforeEach] [Testpattern: Inline-volume (default fs)] volumes
  /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/storage/framework/testsuite.go:51
Jul 18 19:54:39.358: INFO: Driver local doesn't support InlineVolume -- skipping
... skipping 95 lines ...
[BeforeEach] Basic StatefulSet functionality [StatefulSetBasic]
  /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/apps/statefulset.go:105
STEP: Creating service test in namespace statefulset-3849
[It] should not deadlock when a pod's predecessor fails
  /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/apps/statefulset.go:250
STEP: Creating statefulset ss in namespace statefulset-3849
Jul 18 19:54:39.669: INFO: error finding default storageClass : No default storage class found
[AfterEach] Basic StatefulSet functionality [StatefulSetBasic]
  /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/apps/statefulset.go:116
Jul 18 19:54:39.670: INFO: Deleting all statefulset in ns statefulset-3849
[AfterEach] [sig-apps] StatefulSet
  /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:186
Jul 18 19:54:39.761: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready
... skipping 5 lines ...
/workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/apps/framework.go:23
  Basic StatefulSet functionality [StatefulSetBasic]
  /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/apps/statefulset.go:95
    should not deadlock when a pod's predecessor fails [It]
    /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/apps/statefulset.go:250

    error finding default storageClass : No default storage class found

    /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/pv/pv.go:819
------------------------------
SS
------------------------------
[BeforeEach] [sig-storage] EmptyDir volumes
... skipping 2 lines ...
Jul 18 19:54:17.720: INFO: >>> kubeConfig: /root/.kube/config
STEP: Building a namespace api object, basename emptydir
STEP: Waiting for a default service account to be provisioned in namespace
[It] should support (root,0777,default) [LinuxOnly] [NodeConformance] [Conformance]
  /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:630
STEP: Creating a pod to test emptydir 0777 on node default medium
Jul 18 19:54:17.921: INFO: Waiting up to 5m0s for pod "pod-9e2ba5cd-c399-408f-9388-2a5db0c03f76" in namespace "emptydir-8893" to be "Succeeded or Failed"
Jul 18 19:54:17.952: INFO: Pod "pod-9e2ba5cd-c399-408f-9388-2a5db0c03f76": Phase="Pending", Reason="", readiness=false. Elapsed: 30.911784ms
Jul 18 19:54:19.984: INFO: Pod "pod-9e2ba5cd-c399-408f-9388-2a5db0c03f76": Phase="Pending", Reason="", readiness=false. Elapsed: 2.063378058s
Jul 18 19:54:22.018: INFO: Pod "pod-9e2ba5cd-c399-408f-9388-2a5db0c03f76": Phase="Pending", Reason="", readiness=false. Elapsed: 4.096946169s
Jul 18 19:54:24.050: INFO: Pod "pod-9e2ba5cd-c399-408f-9388-2a5db0c03f76": Phase="Pending", Reason="", readiness=false. Elapsed: 6.129101263s
Jul 18 19:54:26.081: INFO: Pod "pod-9e2ba5cd-c399-408f-9388-2a5db0c03f76": Phase="Pending", Reason="", readiness=false. Elapsed: 8.160354858s
Jul 18 19:54:28.113: INFO: Pod "pod-9e2ba5cd-c399-408f-9388-2a5db0c03f76": Phase="Pending", Reason="", readiness=false. Elapsed: 10.192030975s
Jul 18 19:54:30.145: INFO: Pod "pod-9e2ba5cd-c399-408f-9388-2a5db0c03f76": Phase="Pending", Reason="", readiness=false. Elapsed: 12.224139686s
Jul 18 19:54:32.179: INFO: Pod "pod-9e2ba5cd-c399-408f-9388-2a5db0c03f76": Phase="Pending", Reason="", readiness=false. Elapsed: 14.257807995s
Jul 18 19:54:34.212: INFO: Pod "pod-9e2ba5cd-c399-408f-9388-2a5db0c03f76": Phase="Pending", Reason="", readiness=false. Elapsed: 16.290815872s
Jul 18 19:54:36.263: INFO: Pod "pod-9e2ba5cd-c399-408f-9388-2a5db0c03f76": Phase="Pending", Reason="", readiness=false. Elapsed: 18.34205479s
Jul 18 19:54:38.295: INFO: Pod "pod-9e2ba5cd-c399-408f-9388-2a5db0c03f76": Phase="Pending", Reason="", readiness=false. Elapsed: 20.374145011s
Jul 18 19:54:40.329: INFO: Pod "pod-9e2ba5cd-c399-408f-9388-2a5db0c03f76": Phase="Succeeded", Reason="", readiness=false. Elapsed: 22.407906313s
STEP: Saw pod success
Jul 18 19:54:40.329: INFO: Pod "pod-9e2ba5cd-c399-408f-9388-2a5db0c03f76" satisfied condition "Succeeded or Failed"
Jul 18 19:54:40.361: INFO: Trying to get logs from node 10.116.0.7 pod pod-9e2ba5cd-c399-408f-9388-2a5db0c03f76 container test-container: <nil>
STEP: delete the pod
Jul 18 19:54:40.445: INFO: Waiting for pod pod-9e2ba5cd-c399-408f-9388-2a5db0c03f76 to disappear
Jul 18 19:54:40.476: INFO: Pod pod-9e2ba5cd-c399-408f-9388-2a5db0c03f76 no longer exists
[AfterEach] [sig-storage] EmptyDir volumes
  /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:186
... skipping 4 lines ...
• [SLOW TEST:22.853 seconds]
[sig-storage] EmptyDir volumes
/workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/common/storage/framework.go:23
  should support (root,0777,default) [LinuxOnly] [NodeConformance] [Conformance]
  /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:630
------------------------------
{"msg":"PASSED [sig-storage] EmptyDir volumes should support (root,0777,default) [LinuxOnly] [NodeConformance] [Conformance]","total":-1,"completed":9,"skipped":63,"failed":0}
[BeforeEach] [Testpattern: Pre-provisioned PV (ext3)] volumes
  /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/storage/framework/testsuite.go:51
Jul 18 19:54:40.582: INFO: Driver csi-hostpath doesn't support PreprovisionedPV -- skipping
[AfterEach] [Testpattern: Pre-provisioned PV (ext3)] volumes
  /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:186

... skipping 143 lines ...
/workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/node/framework.go:23
  Clean up pods on node
  /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/node/kubelet.go:279
    kubelet should be able to delete 10 pods per node in 1m0s.
    /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/node/kubelet.go:341
------------------------------
{"msg":"PASSED [sig-node] kubelet Clean up pods on node kubelet should be able to delete 10 pods per node in 1m0s.","total":-1,"completed":7,"skipped":32,"failed":0}
[BeforeEach] [Testpattern: Dynamic PV (default fs)] fsgroupchangepolicy
  /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/storage/framework/testsuite.go:51
Jul 18 19:54:40.650: INFO: Driver local doesn't support DynamicPV -- skipping
[AfterEach] [Testpattern: Dynamic PV (default fs)] fsgroupchangepolicy
  /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:186

... skipping 204 lines ...
Jul 18 19:53:05.782: INFO: creating *v1.ClusterRoleBinding: csi-controller-snapshotter-role-csi-mock-volumes-8754
Jul 18 19:53:05.816: INFO: creating *v1.StorageClass: csi-mock-sc-csi-mock-volumes-8754
Jul 18 19:53:05.849: INFO: creating *v1.StatefulSet: csi-mock-volumes-8754-8056/csi-mockplugin
Jul 18 19:53:05.888: INFO: creating *v1.CSIDriver: csi-mock-csi-mock-volumes-8754
Jul 18 19:53:05.922: INFO: waiting up to 4m0s for CSIDriver "csi-mock-csi-mock-volumes-8754"
Jul 18 19:53:05.953: INFO: waiting for CSIDriver csi-mock-csi-mock-volumes-8754 to register on node 10.116.0.5
I0718 19:53:10.381230   12367 csi.go:392] gRPCCall: {"Method":"/csi.v1.Identity/Probe","Request":{},"Response":{"ready":{"value":true}},"Error":"","FullError":null}
I0718 19:53:10.412620   12367 csi.go:392] gRPCCall: {"Method":"/csi.v1.Identity/GetPluginInfo","Request":{},"Response":{"name":"csi-mock-csi-mock-volumes-8754","vendor_version":"0.3.0","manifest":{"url":"https://k8s.io/kubernetes/test/e2e/storage/drivers/csi-test/mock"}},"Error":"","FullError":null}
I0718 19:53:10.442800   12367 csi.go:392] gRPCCall: {"Method":"/csi.v1.Identity/GetPluginCapabilities","Request":{},"Response":{"capabilities":[{"Type":{"Service":{"type":1}}},{"Type":{"VolumeExpansion":{"type":1}}}]},"Error":"","FullError":null}
I0718 19:53:10.477536   12367 csi.go:392] gRPCCall: {"Method":"/csi.v1.Controller/ControllerGetCapabilities","Request":{},"Response":{"capabilities":[{"Type":{"Rpc":{"type":1}}},{"Type":{"Rpc":{"type":3}}},{"Type":{"Rpc":{"type":10}}},{"Type":{"Rpc":{"type":4}}},{"Type":{"Rpc":{"type":6}}},{"Type":{"Rpc":{"type":5}}},{"Type":{"Rpc":{"type":8}}},{"Type":{"Rpc":{"type":7}}},{"Type":{"Rpc":{"type":12}}},{"Type":{"Rpc":{"type":11}}},{"Type":{"Rpc":{"type":9}}}]},"Error":"","FullError":null}
I0718 19:53:10.523355   12367 csi.go:392] gRPCCall: {"Method":"/csi.v1.Identity/GetPluginInfo","Request":{},"Response":{"name":"csi-mock-csi-mock-volumes-8754","vendor_version":"0.3.0","manifest":{"url":"https://k8s.io/kubernetes/test/e2e/storage/drivers/csi-test/mock"}},"Error":"","FullError":null}
I0718 19:53:10.683395   12367 csi.go:392] gRPCCall: {"Method":"/csi.v1.Node/NodeGetInfo","Request":{},"Response":{"node_id":"csi-mock-csi-mock-volumes-8754"},"Error":"","FullError":null}
STEP: Creating pod
Jul 18 19:53:15.648: INFO: Warning: Making PVC: VolumeMode specified as invalid empty string, treating as nil
Jul 18 19:53:15.684: INFO: Waiting up to timeout=5m0s for PersistentVolumeClaims [pvc-mcgvn] to have phase Bound
I0718 19:53:15.698266   12367 csi.go:392] gRPCCall: {"Method":"/csi.v1.Controller/CreateVolume","Request":{"name":"pvc-e87ea409-27d3-456d-b3d6-3b222ddcbb05","capacity_range":{"required_bytes":1073741824},"volume_capabilities":[{"AccessType":{"Mount":{"fs_type":"ext4"}},"access_mode":{"mode":1}}]},"Response":null,"Error":"rpc error: code = ResourceExhausted desc = fake error","FullError":{"code":8,"message":"fake error"}}
Jul 18 19:53:15.715: INFO: PersistentVolumeClaim pvc-mcgvn found but phase is Pending instead of Bound.
I0718 19:53:15.727975   12367 csi.go:392] gRPCCall: {"Method":"/csi.v1.Controller/CreateVolume","Request":{"name":"pvc-e87ea409-27d3-456d-b3d6-3b222ddcbb05","capacity_range":{"required_bytes":1073741824},"volume_capabilities":[{"AccessType":{"Mount":{"fs_type":"ext4"}},"access_mode":{"mode":1}}]},"Response":{"volume":{"capacity_bytes":1073741824,"volume_id":"4","volume_context":{"name":"pvc-e87ea409-27d3-456d-b3d6-3b222ddcbb05"}}},"Error":"","FullError":null}
Jul 18 19:53:17.747: INFO: PersistentVolumeClaim pvc-mcgvn found and phase=Bound (2.06321511s)
I0718 19:53:19.781326   12367 csi.go:392] gRPCCall: {"Method":"/csi.v1.Node/NodeGetCapabilities","Request":{},"Response":{"capabilities":[{"Type":{"Rpc":{}}},{"Type":{"Rpc":{"type":1}}},{"Type":{"Rpc":{"type":2}}},{"Type":{"Rpc":{"type":4}}}]},"Error":"","FullError":null}
Jul 18 19:53:19.811: INFO: >>> kubeConfig: /root/.kube/config
I0718 19:53:20.058172   12367 csi.go:392] gRPCCall: {"Method":"/csi.v1.Node/NodeStageVolume","Request":{"volume_id":"4","staging_target_path":"/var/lib/kubelet/plugins/kubernetes.io/csi/pv/pvc-e87ea409-27d3-456d-b3d6-3b222ddcbb05/globalmount","volume_capability":{"AccessType":{"Mount":{"fs_type":"ext4"}},"access_mode":{"mode":1}},"volume_context":{"name":"pvc-e87ea409-27d3-456d-b3d6-3b222ddcbb05","storage.kubernetes.io/csiProvisionerIdentity":"1626637990490-8081-csi-mock-csi-mock-volumes-8754"}},"Response":{},"Error":"","FullError":null}
I0718 19:53:21.036209   12367 csi.go:392] gRPCCall: {"Method":"/csi.v1.Node/NodeGetCapabilities","Request":{},"Response":{"capabilities":[{"Type":{"Rpc":{}}},{"Type":{"Rpc":{"type":1}}},{"Type":{"Rpc":{"type":2}}},{"Type":{"Rpc":{"type":4}}}]},"Error":"","FullError":null}
Jul 18 19:53:21.065: INFO: >>> kubeConfig: /root/.kube/config
Jul 18 19:53:21.309: INFO: >>> kubeConfig: /root/.kube/config
Jul 18 19:53:21.561: INFO: >>> kubeConfig: /root/.kube/config
I0718 19:53:21.817261   12367 csi.go:392] gRPCCall: {"Method":"/csi.v1.Node/NodePublishVolume","Request":{"volume_id":"4","staging_target_path":"/var/lib/kubelet/plugins/kubernetes.io/csi/pv/pvc-e87ea409-27d3-456d-b3d6-3b222ddcbb05/globalmount","target_path":"/var/lib/kubelet/pods/deaf3613-1685-43e8-90dd-356290c20e9b/volumes/kubernetes.io~csi/pvc-e87ea409-27d3-456d-b3d6-3b222ddcbb05/mount","volume_capability":{"AccessType":{"Mount":{"fs_type":"ext4"}},"access_mode":{"mode":1}},"volume_context":{"name":"pvc-e87ea409-27d3-456d-b3d6-3b222ddcbb05","storage.kubernetes.io/csiProvisionerIdentity":"1626637990490-8081-csi-mock-csi-mock-volumes-8754"}},"Response":{},"Error":"","FullError":null}
Jul 18 19:53:31.908: INFO: Deleting pod "pvc-volume-tester-7vgkb" in namespace "csi-mock-volumes-8754"
Jul 18 19:53:31.944: INFO: Wait up to 5m0s for pod "pvc-volume-tester-7vgkb" to be fully deleted
Jul 18 19:53:33.281: INFO: >>> kubeConfig: /root/.kube/config
I0718 19:53:33.520058   12367 csi.go:392] gRPCCall: {"Method":"/csi.v1.Node/NodeUnpublishVolume","Request":{"volume_id":"4","target_path":"/var/lib/kubelet/pods/deaf3613-1685-43e8-90dd-356290c20e9b/volumes/kubernetes.io~csi/pvc-e87ea409-27d3-456d-b3d6-3b222ddcbb05/mount"},"Response":{},"Error":"","FullError":null}
I0718 19:53:33.587137   12367 csi.go:392] gRPCCall: {"Method":"/csi.v1.Node/NodeGetCapabilities","Request":{},"Response":{"capabilities":[{"Type":{"Rpc":{}}},{"Type":{"Rpc":{"type":1}}},{"Type":{"Rpc":{"type":2}}},{"Type":{"Rpc":{"type":4}}}]},"Error":"","FullError":null}
I0718 19:53:33.617032   12367 csi.go:392] gRPCCall: {"Method":"/csi.v1.Node/NodeUnstageVolume","Request":{"volume_id":"4","staging_target_path":"/var/lib/kubelet/plugins/kubernetes.io/csi/pv/pvc-e87ea409-27d3-456d-b3d6-3b222ddcbb05/globalmount"},"Response":{},"Error":"","FullError":null}
I0718 19:53:38.069639   12367 csi.go:392] gRPCCall: {"Method":"/csi.v1.Controller/DeleteVolume","Request":{"volume_id":"4"},"Response":{},"Error":"","FullError":null}
STEP: Checking PVC events
Jul 18 19:53:39.042: INFO: PVC event ADDED: &v1.PersistentVolumeClaim{TypeMeta:v1.TypeMeta{Kind:"", APIVersion:""}, ObjectMeta:v1.ObjectMeta{Name:"pvc-mcgvn", GenerateName:"pvc-", Namespace:"csi-mock-volumes-8754", SelfLink:"", UID:"e87ea409-27d3-456d-b3d6-3b222ddcbb05", ResourceVersion:"8794", Generation:0, CreationTimestamp:v1.Time{Time:time.Time{wall:0x0, ext:63762234795, loc:(*time.Location)(0x9ddf5a0)}}, DeletionTimestamp:(*v1.Time)(nil), DeletionGracePeriodSeconds:(*int64)(nil), Labels:map[string]string(nil), Annotations:map[string]string(nil), OwnerReferences:[]v1.OwnerReference(nil), Finalizers:[]string{"kubernetes.io/pvc-protection"}, ClusterName:"", ManagedFields:[]v1.ManagedFieldsEntry{v1.ManagedFieldsEntry{Manager:"e2e.test", Operation:"Update", APIVersion:"v1", Time:(*v1.Time)(0xc002bc3398), FieldsType:"FieldsV1", FieldsV1:(*v1.FieldsV1)(0xc002bc33b0)}}}, Spec:v1.PersistentVolumeClaimSpec{AccessModes:[]v1.PersistentVolumeAccessMode{"ReadWriteOnce"}, Selector:(*v1.LabelSelector)(nil), Resources:v1.ResourceRequirements{Limits:v1.ResourceList(nil), Requests:v1.ResourceList{"storage":resource.Quantity{i:resource.int64Amount{value:1073741824, scale:0}, d:resource.infDecAmount{Dec:(*inf.Dec)(nil)}, s:"1Gi", Format:"BinarySI"}}}, VolumeName:"", StorageClassName:(*string)(0xc00294def0), VolumeMode:(*v1.PersistentVolumeMode)(0xc00294df00), DataSource:(*v1.TypedLocalObjectReference)(nil)}, Status:v1.PersistentVolumeClaimStatus{Phase:"Pending", AccessModes:[]v1.PersistentVolumeAccessMode(nil), Capacity:v1.ResourceList(nil), Conditions:[]v1.PersistentVolumeClaimCondition(nil)}}
Jul 18 19:53:39.043: INFO: PVC event MODIFIED: &v1.PersistentVolumeClaim{TypeMeta:v1.TypeMeta{Kind:"", APIVersion:""}, ObjectMeta:v1.ObjectMeta{Name:"pvc-mcgvn", GenerateName:"pvc-", Namespace:"csi-mock-volumes-8754", SelfLink:"", UID:"e87ea409-27d3-456d-b3d6-3b222ddcbb05", ResourceVersion:"8796", Generation:0, CreationTimestamp:v1.Time{Time:time.Time{wall:0x0, ext:63762234795, loc:(*time.Location)(0x9ddf5a0)}}, DeletionTimestamp:(*v1.Time)(nil), DeletionGracePeriodSeconds:(*int64)(nil), Labels:map[string]string(nil), Annotations:map[string]string{"volume.beta.kubernetes.io/storage-provisioner":"csi-mock-csi-mock-volumes-8754"}, OwnerReferences:[]v1.OwnerReference(nil), Finalizers:[]string{"kubernetes.io/pvc-protection"}, ClusterName:"", ManagedFields:[]v1.ManagedFieldsEntry{v1.ManagedFieldsEntry{Manager:"e2e.test", Operation:"Update", APIVersion:"v1", Time:(*v1.Time)(0xc002b9dc08), FieldsType:"FieldsV1", FieldsV1:(*v1.FieldsV1)(0xc002b9dc20)}, v1.ManagedFieldsEntry{Manager:"kube-controller-manager", Operation:"Update", APIVersion:"v1", Time:(*v1.Time)(0xc002b9dc38), FieldsType:"FieldsV1", FieldsV1:(*v1.FieldsV1)(0xc002b9dc50)}}}, Spec:v1.PersistentVolumeClaimSpec{AccessModes:[]v1.PersistentVolumeAccessMode{"ReadWriteOnce"}, Selector:(*v1.LabelSelector)(nil), Resources:v1.ResourceRequirements{Limits:v1.ResourceList(nil), Requests:v1.ResourceList{"storage":resource.Quantity{i:resource.int64Amount{value:1073741824, scale:0}, d:resource.infDecAmount{Dec:(*inf.Dec)(nil)}, s:"1Gi", Format:"BinarySI"}}}, VolumeName:"", StorageClassName:(*string)(0xc0028eda30), VolumeMode:(*v1.PersistentVolumeMode)(0xc0028eda40), DataSource:(*v1.TypedLocalObjectReference)(nil)}, Status:v1.PersistentVolumeClaimStatus{Phase:"Pending", AccessModes:[]v1.PersistentVolumeAccessMode(nil), Capacity:v1.ResourceList(nil), Conditions:[]v1.PersistentVolumeClaimCondition(nil)}}
Jul 18 19:53:39.043: INFO: PVC event MODIFIED: &v1.PersistentVolumeClaim{TypeMeta:v1.TypeMeta{Kind:"", APIVersion:""}, ObjectMeta:v1.ObjectMeta{Name:"pvc-mcgvn", GenerateName:"pvc-", Namespace:"csi-mock-volumes-8754", SelfLink:"", UID:"e87ea409-27d3-456d-b3d6-3b222ddcbb05", ResourceVersion:"8802", Generation:0, CreationTimestamp:v1.Time{Time:time.Time{wall:0x0, ext:63762234795, loc:(*time.Location)(0x9ddf5a0)}}, DeletionTimestamp:(*v1.Time)(nil), DeletionGracePeriodSeconds:(*int64)(nil), Labels:map[string]string(nil), Annotations:map[string]string{"pv.kubernetes.io/bind-completed":"yes", "pv.kubernetes.io/bound-by-controller":"yes", "volume.beta.kubernetes.io/storage-provisioner":"csi-mock-csi-mock-volumes-8754"}, OwnerReferences:[]v1.OwnerReference(nil), Finalizers:[]string{"kubernetes.io/pvc-protection"}, ClusterName:"", ManagedFields:[]v1.ManagedFieldsEntry{v1.ManagedFieldsEntry{Manager:"e2e.test", Operation:"Update", APIVersion:"v1", Time:(*v1.Time)(0xc002695638), FieldsType:"FieldsV1", FieldsV1:(*v1.FieldsV1)(0xc002695650)}, v1.ManagedFieldsEntry{Manager:"kube-controller-manager", Operation:"Update", APIVersion:"v1", Time:(*v1.Time)(0xc002695668), FieldsType:"FieldsV1", FieldsV1:(*v1.FieldsV1)(0xc002695680)}}}, Spec:v1.PersistentVolumeClaimSpec{AccessModes:[]v1.PersistentVolumeAccessMode{"ReadWriteOnce"}, Selector:(*v1.LabelSelector)(nil), Resources:v1.ResourceRequirements{Limits:v1.ResourceList(nil), Requests:v1.ResourceList{"storage":resource.Quantity{i:resource.int64Amount{value:1073741824, scale:0}, d:resource.infDecAmount{Dec:(*inf.Dec)(nil)}, s:"1Gi", Format:"BinarySI"}}}, VolumeName:"pvc-e87ea409-27d3-456d-b3d6-3b222ddcbb05", StorageClassName:(*string)(0xc003230c90), VolumeMode:(*v1.PersistentVolumeMode)(0xc003230ca0), DataSource:(*v1.TypedLocalObjectReference)(nil)}, Status:v1.PersistentVolumeClaimStatus{Phase:"Pending", AccessModes:[]v1.PersistentVolumeAccessMode(nil), Capacity:v1.ResourceList(nil), Conditions:[]v1.PersistentVolumeClaimCondition(nil)}}
Jul 18 19:53:39.043: INFO: PVC event MODIFIED: &v1.PersistentVolumeClaim{TypeMeta:v1.TypeMeta{Kind:"", APIVersion:""}, ObjectMeta:v1.ObjectMeta{Name:"pvc-mcgvn", GenerateName:"pvc-", Namespace:"csi-mock-volumes-8754", SelfLink:"", UID:"e87ea409-27d3-456d-b3d6-3b222ddcbb05", ResourceVersion:"8803", Generation:0, CreationTimestamp:v1.Time{Time:time.Time{wall:0x0, ext:63762234795, loc:(*time.Location)(0x9ddf5a0)}}, DeletionTimestamp:(*v1.Time)(nil), DeletionGracePeriodSeconds:(*int64)(nil), Labels:map[string]string(nil), Annotations:map[string]string{"pv.kubernetes.io/bind-completed":"yes", "pv.kubernetes.io/bound-by-controller":"yes", "volume.beta.kubernetes.io/storage-provisioner":"csi-mock-csi-mock-volumes-8754"}, OwnerReferences:[]v1.OwnerReference(nil), Finalizers:[]string{"kubernetes.io/pvc-protection"}, ClusterName:"", ManagedFields:[]v1.ManagedFieldsEntry{v1.ManagedFieldsEntry{Manager:"e2e.test", Operation:"Update", APIVersion:"v1", Time:(*v1.Time)(0xc0026956b0), FieldsType:"FieldsV1", FieldsV1:(*v1.FieldsV1)(0xc0026956c8)}, v1.ManagedFieldsEntry{Manager:"kube-controller-manager", Operation:"Update", APIVersion:"v1", Time:(*v1.Time)(0xc0026956e0), FieldsType:"FieldsV1", FieldsV1:(*v1.FieldsV1)(0xc0026956f8)}}}, Spec:v1.PersistentVolumeClaimSpec{AccessModes:[]v1.PersistentVolumeAccessMode{"ReadWriteOnce"}, Selector:(*v1.LabelSelector)(nil), Resources:v1.ResourceRequirements{Limits:v1.ResourceList(nil), Requests:v1.ResourceList{"storage":resource.Quantity{i:resource.int64Amount{value:1073741824, scale:0}, d:resource.infDecAmount{Dec:(*inf.Dec)(nil)}, s:"1Gi", Format:"BinarySI"}}}, VolumeName:"pvc-e87ea409-27d3-456d-b3d6-3b222ddcbb05", StorageClassName:(*string)(0xc003230cd0), VolumeMode:(*v1.PersistentVolumeMode)(0xc003230ce0), DataSource:(*v1.TypedLocalObjectReference)(nil)}, Status:v1.PersistentVolumeClaimStatus{Phase:"Bound", AccessModes:[]v1.PersistentVolumeAccessMode{"ReadWriteOnce"}, Capacity:v1.ResourceList{"storage":resource.Quantity{i:resource.int64Amount{value:1073741824, scale:0}, d:resource.infDecAmount{Dec:(*inf.Dec)(nil)}, s:"1Gi", Format:"BinarySI"}}, Conditions:[]v1.PersistentVolumeClaimCondition(nil)}}
Jul 18 19:53:39.043: INFO: PVC event MODIFIED: &v1.PersistentVolumeClaim{TypeMeta:v1.TypeMeta{Kind:"", APIVersion:""}, ObjectMeta:v1.ObjectMeta{Name:"pvc-mcgvn", GenerateName:"pvc-", Namespace:"csi-mock-volumes-8754", SelfLink:"", UID:"e87ea409-27d3-456d-b3d6-3b222ddcbb05", ResourceVersion:"9788", Generation:0, CreationTimestamp:v1.Time{Time:time.Time{wall:0x0, ext:63762234795, loc:(*time.Location)(0x9ddf5a0)}}, DeletionTimestamp:(*v1.Time)(0xc002695728), DeletionGracePeriodSeconds:(*int64)(0xc0001926f8), Labels:map[string]string(nil), Annotations:map[string]string{"pv.kubernetes.io/bind-completed":"yes", "pv.kubernetes.io/bound-by-controller":"yes", "volume.beta.kubernetes.io/storage-provisioner":"csi-mock-csi-mock-volumes-8754"}, OwnerReferences:[]v1.OwnerReference(nil), Finalizers:[]string{"kubernetes.io/pvc-protection"}, ClusterName:"", ManagedFields:[]v1.ManagedFieldsEntry{v1.ManagedFieldsEntry{Manager:"e2e.test", Operation:"Update", APIVersion:"v1", Time:(*v1.Time)(0xc002695740), FieldsType:"FieldsV1", FieldsV1:(*v1.FieldsV1)(0xc002695758)}, v1.ManagedFieldsEntry{Manager:"kube-controller-manager", Operation:"Update", APIVersion:"v1", Time:(*v1.Time)(0xc002695770), FieldsType:"FieldsV1", FieldsV1:(*v1.FieldsV1)(0xc002695788)}}}, Spec:v1.PersistentVolumeClaimSpec{AccessModes:[]v1.PersistentVolumeAccessMode{"ReadWriteOnce"}, Selector:(*v1.LabelSelector)(nil), Resources:v1.ResourceRequirements{Limits:v1.ResourceList(nil), Requests:v1.ResourceList{"storage":resource.Quantity{i:resource.int64Amount{value:1073741824, scale:0}, d:resource.infDecAmount{Dec:(*inf.Dec)(nil)}, s:"1Gi", Format:"BinarySI"}}}, VolumeName:"pvc-e87ea409-27d3-456d-b3d6-3b222ddcbb05", StorageClassName:(*string)(0xc003230d20), VolumeMode:(*v1.PersistentVolumeMode)(0xc003230d30), DataSource:(*v1.TypedLocalObjectReference)(nil)}, Status:v1.PersistentVolumeClaimStatus{Phase:"Bound", AccessModes:[]v1.PersistentVolumeAccessMode{"ReadWriteOnce"}, Capacity:v1.ResourceList{"storage":resource.Quantity{i:resource.int64Amount{value:1073741824, scale:0}, d:resource.infDecAmount{Dec:(*inf.Dec)(nil)}, s:"1Gi", Format:"BinarySI"}}, Conditions:[]v1.PersistentVolumeClaimCondition(nil)}}
... skipping 48 lines ...
/workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/storage/utils/framework.go:23
  storage capacity
  /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/storage/csi_mock_volume.go:900
    exhausted, immediate binding
    /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/storage/csi_mock_volume.go:958
------------------------------
{"msg":"PASSED [sig-storage] CSI mock volume storage capacity exhausted, immediate binding","total":-1,"completed":3,"skipped":12,"failed":0}
[BeforeEach] [Testpattern: Inline-volume (default fs)] volumes
  /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/storage/framework/testsuite.go:51
Jul 18 19:54:41.687: INFO: Only supported for node OS distro [gci ubuntu custom] (not debian)
[AfterEach] [Testpattern: Inline-volume (default fs)] volumes
  /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:186

... skipping 107 lines ...
• [SLOW TEST:32.724 seconds]
[sig-apps] Deployment
/workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/apps/framework.go:23
  RollingUpdateDeployment should delete old pods and create new ones [Conformance]
  /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:630
------------------------------
{"msg":"PASSED [sig-apps] Deployment RollingUpdateDeployment should delete old pods and create new ones [Conformance]","total":-1,"completed":6,"skipped":50,"failed":0}
[BeforeEach] [Testpattern: Inline-volume (ext4)] volumes
  /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/storage/framework/testsuite.go:51
Jul 18 19:54:42.879: INFO: Only supported for providers [gce gke] (not skeleton)
[AfterEach] [Testpattern: Inline-volume (ext4)] volumes
  /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:186

... skipping 88 lines ...
• [SLOW TEST:94.003 seconds]
[sig-storage] ConfigMap
/workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/common/storage/framework.go:23
  updates should be reflected in volume [NodeConformance] [Conformance]
  /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:630
------------------------------
{"msg":"PASSED [sig-storage] ConfigMap updates should be reflected in volume [NodeConformance] [Conformance]","total":-1,"completed":7,"skipped":73,"failed":0}
[BeforeEach] [Testpattern: Inline-volume (default fs)] subPath
  /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/storage/framework/testsuite.go:51
Jul 18 19:54:46.297: INFO: Only supported for providers [gce gke] (not skeleton)
[AfterEach] [Testpattern: Inline-volume (default fs)] subPath
  /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:186

... skipping 82 lines ...
  /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/storage/persistent_volumes-local.go:192
    Two pods mounting a local volume one after the other
    /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/storage/persistent_volumes-local.go:254
      should be able to write from pod1 and read from pod2
      /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/storage/persistent_volumes-local.go:255
------------------------------
{"msg":"PASSED [sig-storage] PersistentVolumes-local  [Volume type: tmpfs] Two pods mounting a local volume one after the other should be able to write from pod1 and read from pod2","total":-1,"completed":10,"skipped":74,"failed":0}
[BeforeEach] [Testpattern: Pre-provisioned PV (default fs)] subPath
  /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/storage/framework/testsuite.go:51
Jul 18 19:54:47.046: INFO: Only supported for node OS distro [gci ubuntu custom] (not debian)
[AfterEach] [Testpattern: Pre-provisioned PV (default fs)] subPath
  /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:186

... skipping 34 lines ...
Jul 18 19:54:47.292: INFO: pv is nil


S [SKIPPING] in Spec Setup (BeforeEach) [0.231 seconds]
[sig-storage] PersistentVolumes GCEPD
/workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/storage/utils/framework.go:23
  should test that deleting the PV before the pod does not cause pod deletion to fail on PD detach [BeforeEach]
  /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/storage/persistent_volumes-gce.go:142

  Only supported for providers [gce gke] (not skeleton)

  /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/storage/persistent_volumes-gce.go:85
------------------------------
... skipping 117 lines ...
STEP: Building a namespace api object, basename projected
STEP: Waiting for a default service account to be provisioned in namespace
[It] should be consumable from pods in volume with mappings as non-root with FSGroup [LinuxOnly] [NodeFeature:FSGroup]
  /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/common/storage/projected_configmap.go:110
STEP: Creating configMap with name projected-configmap-test-volume-map-cae08a56-1f71-437d-9b33-94efe773579f
STEP: Creating a pod to test consume configMaps
Jul 18 19:54:27.423: INFO: Waiting up to 5m0s for pod "pod-projected-configmaps-043799c9-a483-4eb2-9c5f-b053970e8ecc" in namespace "projected-3199" to be "Succeeded or Failed"
Jul 18 19:54:27.454: INFO: Pod "pod-projected-configmaps-043799c9-a483-4eb2-9c5f-b053970e8ecc": Phase="Pending", Reason="", readiness=false. Elapsed: 30.980644ms
Jul 18 19:54:29.487: INFO: Pod "pod-projected-configmaps-043799c9-a483-4eb2-9c5f-b053970e8ecc": Phase="Pending", Reason="", readiness=false. Elapsed: 2.063328543s
Jul 18 19:54:31.520: INFO: Pod "pod-projected-configmaps-043799c9-a483-4eb2-9c5f-b053970e8ecc": Phase="Pending", Reason="", readiness=false. Elapsed: 4.09609065s
Jul 18 19:54:33.552: INFO: Pod "pod-projected-configmaps-043799c9-a483-4eb2-9c5f-b053970e8ecc": Phase="Pending", Reason="", readiness=false. Elapsed: 6.128481778s
Jul 18 19:54:35.583: INFO: Pod "pod-projected-configmaps-043799c9-a483-4eb2-9c5f-b053970e8ecc": Phase="Pending", Reason="", readiness=false. Elapsed: 8.159937055s
Jul 18 19:54:37.616: INFO: Pod "pod-projected-configmaps-043799c9-a483-4eb2-9c5f-b053970e8ecc": Phase="Pending", Reason="", readiness=false. Elapsed: 10.192150824s
Jul 18 19:54:39.647: INFO: Pod "pod-projected-configmaps-043799c9-a483-4eb2-9c5f-b053970e8ecc": Phase="Pending", Reason="", readiness=false. Elapsed: 12.223738757s
Jul 18 19:54:41.680: INFO: Pod "pod-projected-configmaps-043799c9-a483-4eb2-9c5f-b053970e8ecc": Phase="Pending", Reason="", readiness=false. Elapsed: 14.256446444s
Jul 18 19:54:43.712: INFO: Pod "pod-projected-configmaps-043799c9-a483-4eb2-9c5f-b053970e8ecc": Phase="Pending", Reason="", readiness=false. Elapsed: 16.288334069s
Jul 18 19:54:45.751: INFO: Pod "pod-projected-configmaps-043799c9-a483-4eb2-9c5f-b053970e8ecc": Phase="Pending", Reason="", readiness=false. Elapsed: 18.327945803s
Jul 18 19:54:47.782: INFO: Pod "pod-projected-configmaps-043799c9-a483-4eb2-9c5f-b053970e8ecc": Phase="Succeeded", Reason="", readiness=false. Elapsed: 20.358936855s
STEP: Saw pod success
Jul 18 19:54:47.782: INFO: Pod "pod-projected-configmaps-043799c9-a483-4eb2-9c5f-b053970e8ecc" satisfied condition "Succeeded or Failed"
Jul 18 19:54:47.814: INFO: Trying to get logs from node 10.116.0.5 pod pod-projected-configmaps-043799c9-a483-4eb2-9c5f-b053970e8ecc container agnhost-container: <nil>
STEP: delete the pod
Jul 18 19:54:47.896: INFO: Waiting for pod pod-projected-configmaps-043799c9-a483-4eb2-9c5f-b053970e8ecc to disappear
Jul 18 19:54:47.926: INFO: Pod pod-projected-configmaps-043799c9-a483-4eb2-9c5f-b053970e8ecc no longer exists
[AfterEach] [sig-storage] Projected configMap
  /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:186
... skipping 4 lines ...
• [SLOW TEST:20.825 seconds]
[sig-storage] Projected configMap
/workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/common/storage/framework.go:23
  should be consumable from pods in volume with mappings as non-root with FSGroup [LinuxOnly] [NodeFeature:FSGroup]
  /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/common/storage/projected_configmap.go:110
------------------------------
{"msg":"PASSED [sig-storage] Projected configMap should be consumable from pods in volume with mappings as non-root with FSGroup [LinuxOnly] [NodeFeature:FSGroup]","total":-1,"completed":12,"skipped":62,"failed":0}

SSSS
------------------------------
[BeforeEach] [Testpattern: Inline-volume (default fs)] volumes
  /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/storage/framework/testsuite.go:51
Jul 18 19:54:48.045: INFO: Driver local doesn't support InlineVolume -- skipping
... skipping 112 lines ...
STEP: Building a namespace api object, basename projected
STEP: Waiting for a default service account to be provisioned in namespace
[It] should be consumable from pods in volume with mappings as non-root [NodeConformance] [Conformance]
  /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:630
STEP: Creating configMap with name projected-configmap-test-volume-map-d273fdea-e434-4410-a9cd-89780de36ccc
STEP: Creating a pod to test consume configMaps
Jul 18 19:54:32.101: INFO: Waiting up to 5m0s for pod "pod-projected-configmaps-b340d8e0-7cc8-4d03-b11f-aad79e9e5ab6" in namespace "projected-3731" to be "Succeeded or Failed"
Jul 18 19:54:32.133: INFO: Pod "pod-projected-configmaps-b340d8e0-7cc8-4d03-b11f-aad79e9e5ab6": Phase="Pending", Reason="", readiness=false. Elapsed: 31.650726ms
Jul 18 19:54:34.165: INFO: Pod "pod-projected-configmaps-b340d8e0-7cc8-4d03-b11f-aad79e9e5ab6": Phase="Pending", Reason="", readiness=false. Elapsed: 2.064069886s
Jul 18 19:54:36.196: INFO: Pod "pod-projected-configmaps-b340d8e0-7cc8-4d03-b11f-aad79e9e5ab6": Phase="Pending", Reason="", readiness=false. Elapsed: 4.095385526s
Jul 18 19:54:38.227: INFO: Pod "pod-projected-configmaps-b340d8e0-7cc8-4d03-b11f-aad79e9e5ab6": Phase="Pending", Reason="", readiness=false. Elapsed: 6.126473655s
Jul 18 19:54:40.258: INFO: Pod "pod-projected-configmaps-b340d8e0-7cc8-4d03-b11f-aad79e9e5ab6": Phase="Pending", Reason="", readiness=false. Elapsed: 8.157513377s
Jul 18 19:54:42.291: INFO: Pod "pod-projected-configmaps-b340d8e0-7cc8-4d03-b11f-aad79e9e5ab6": Phase="Pending", Reason="", readiness=false. Elapsed: 10.189609618s
Jul 18 19:54:44.323: INFO: Pod "pod-projected-configmaps-b340d8e0-7cc8-4d03-b11f-aad79e9e5ab6": Phase="Pending", Reason="", readiness=false. Elapsed: 12.221670978s
Jul 18 19:54:46.357: INFO: Pod "pod-projected-configmaps-b340d8e0-7cc8-4d03-b11f-aad79e9e5ab6": Phase="Pending", Reason="", readiness=false. Elapsed: 14.256006839s
Jul 18 19:54:48.389: INFO: Pod "pod-projected-configmaps-b340d8e0-7cc8-4d03-b11f-aad79e9e5ab6": Phase="Succeeded", Reason="", readiness=false. Elapsed: 16.287751906s
STEP: Saw pod success
Jul 18 19:54:48.389: INFO: Pod "pod-projected-configmaps-b340d8e0-7cc8-4d03-b11f-aad79e9e5ab6" satisfied condition "Succeeded or Failed"
Jul 18 19:54:48.420: INFO: Trying to get logs from node 10.116.0.5 pod pod-projected-configmaps-b340d8e0-7cc8-4d03-b11f-aad79e9e5ab6 container agnhost-container: <nil>
STEP: delete the pod
Jul 18 19:54:48.494: INFO: Waiting for pod pod-projected-configmaps-b340d8e0-7cc8-4d03-b11f-aad79e9e5ab6 to disappear
Jul 18 19:54:48.525: INFO: Pod pod-projected-configmaps-b340d8e0-7cc8-4d03-b11f-aad79e9e5ab6 no longer exists
[AfterEach] [sig-storage] Projected configMap
  /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:186
... skipping 4 lines ...
• [SLOW TEST:16.758 seconds]
[sig-storage] Projected configMap
/workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/common/storage/framework.go:23
  should be consumable from pods in volume with mappings as non-root [NodeConformance] [Conformance]
  /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:630
------------------------------
{"msg":"PASSED [sig-auth] Certificates API [Privileged:ClusterAdmin] should support CSR API operations [Conformance]","total":-1,"completed":6,"skipped":37,"failed":0}
[BeforeEach] [sig-storage] CSI mock volume
  /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:185
STEP: Creating a kubernetes client
Jul 18 19:53:13.517: INFO: >>> kubeConfig: /root/.kube/config
STEP: Building a namespace api object, basename csi-mock-volumes
STEP: Waiting for a default service account to be provisioned in namespace
... skipping 109 lines ...
/workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/storage/utils/framework.go:23
  CSI Volume expansion
  /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/storage/csi_mock_volume.go:561
    should expand volume by restarting pod if attach=off, nodeExpansion=on
    /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/storage/csi_mock_volume.go:590
------------------------------
{"msg":"PASSED [sig-storage] CSI mock volume CSI Volume expansion should expand volume by restarting pod if attach=off, nodeExpansion=on","total":-1,"completed":7,"skipped":37,"failed":0}

S
------------------------------
[BeforeEach] [Testpattern: Dynamic PV (ext3)] volumes
  /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/storage/framework/testsuite.go:51
Jul 18 19:54:48.933: INFO: Driver local doesn't support DynamicPV -- skipping
... skipping 502 lines ...
/workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/kubectl/framework.go:23
  Kubectl logs
  /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/kubectl/kubectl.go:1383
    should be able to retrieve and filter logs  [Conformance]
    /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:630
------------------------------
{"msg":"PASSED [sig-cli] Kubectl client Kubectl logs should be able to retrieve and filter logs  [Conformance]","total":-1,"completed":9,"skipped":74,"failed":0}

SS
------------------------------
[BeforeEach] [sig-storage] CSI mock volume
  /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:185
STEP: Creating a kubernetes client
... skipping 137 lines ...
/workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/kubectl/framework.go:23
  With a server listening on localhost
  /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/kubectl/portforward.go:474
    should support forwarding over websockets
    /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/kubectl/portforward.go:490
------------------------------
{"msg":"PASSED [sig-cli] Kubectl Port forwarding With a server listening on localhost should support forwarding over websockets","total":-1,"completed":8,"skipped":63,"failed":0}

SS
------------------------------
{"msg":"PASSED [sig-storage] Projected configMap should be consumable from pods in volume with mappings as non-root [NodeConformance] [Conformance]","total":-1,"completed":6,"skipped":23,"failed":0}
[BeforeEach] [sig-scheduling] LimitRange
  /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:185
STEP: Creating a kubernetes client
Jul 18 19:54:48.634: INFO: >>> kubeConfig: /root/.kube/config
STEP: Building a namespace api object, basename limitrange
STEP: Waiting for a default service account to be provisioned in namespace
... skipping 34 lines ...
• [SLOW TEST:7.797 seconds]
[sig-scheduling] LimitRange
/workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/scheduling/framework.go:40
  should create a LimitRange with defaults and ensure pod has those defaults applied. [Conformance]
  /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:630
------------------------------
{"msg":"PASSED [sig-scheduling] LimitRange should create a LimitRange with defaults and ensure pod has those defaults applied. [Conformance]","total":-1,"completed":7,"skipped":23,"failed":0}

SS
------------------------------
[BeforeEach] [Testpattern: Dynamic PV (filesystem volmode)] volumeMode
  /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/storage/framework/testsuite.go:51
Jul 18 19:54:56.448: INFO: Driver emptydir doesn't support DynamicPV -- skipping
... skipping 39 lines ...
• [SLOW TEST:16.587 seconds]
[sig-storage] EmptyDir volumes
/workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/common/storage/framework.go:23
  pod should support shared volumes between containers [Conformance]
  /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:630
------------------------------
{"msg":"PASSED [sig-storage] EmptyDir volumes pod should support shared volumes between containers [Conformance]","total":-1,"completed":7,"skipped":62,"failed":0}

SSSSS
------------------------------
[BeforeEach] [Testpattern: Dynamic PV (default fs)(allowExpansion)] volume-expand
  /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/storage/framework/testsuite.go:51
Jul 18 19:54:59.555: INFO: Driver hostPathSymlink doesn't support DynamicPV -- skipping
... skipping 44 lines ...
Jul 18 19:54:40.668: INFO: >>> kubeConfig: /root/.kube/config
STEP: Building a namespace api object, basename emptydir
STEP: Waiting for a default service account to be provisioned in namespace
[It] should support (root,0666,tmpfs) [LinuxOnly] [NodeConformance] [Conformance]
  /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:630
STEP: Creating a pod to test emptydir 0666 on tmpfs
Jul 18 19:54:40.857: INFO: Waiting up to 5m0s for pod "pod-db261340-640f-482a-8c8f-4ccd9a736f61" in namespace "emptydir-6298" to be "Succeeded or Failed"
Jul 18 19:54:40.889: INFO: Pod "pod-db261340-640f-482a-8c8f-4ccd9a736f61": Phase="Pending", Reason="", readiness=false. Elapsed: 31.695983ms
Jul 18 19:54:42.921: INFO: Pod "pod-db261340-640f-482a-8c8f-4ccd9a736f61": Phase="Pending", Reason="", readiness=false. Elapsed: 2.063459877s
Jul 18 19:54:44.953: INFO: Pod "pod-db261340-640f-482a-8c8f-4ccd9a736f61": Phase="Pending", Reason="", readiness=false. Elapsed: 4.096156429s
Jul 18 19:54:46.984: INFO: Pod "pod-db261340-640f-482a-8c8f-4ccd9a736f61": Phase="Pending", Reason="", readiness=false. Elapsed: 6.127355535s
Jul 18 19:54:49.019: INFO: Pod "pod-db261340-640f-482a-8c8f-4ccd9a736f61": Phase="Pending", Reason="", readiness=false. Elapsed: 8.162292755s
Jul 18 19:54:51.051: INFO: Pod "pod-db261340-640f-482a-8c8f-4ccd9a736f61": Phase="Pending", Reason="", readiness=false. Elapsed: 10.19428049s
Jul 18 19:54:53.083: INFO: Pod "pod-db261340-640f-482a-8c8f-4ccd9a736f61": Phase="Pending", Reason="", readiness=false. Elapsed: 12.225775875s
Jul 18 19:54:55.114: INFO: Pod "pod-db261340-640f-482a-8c8f-4ccd9a736f61": Phase="Pending", Reason="", readiness=false. Elapsed: 14.257413933s
Jul 18 19:54:57.199: INFO: Pod "pod-db261340-640f-482a-8c8f-4ccd9a736f61": Phase="Pending", Reason="", readiness=false. Elapsed: 16.341923851s
Jul 18 19:54:59.230: INFO: Pod "pod-db261340-640f-482a-8c8f-4ccd9a736f61": Phase="Pending", Reason="", readiness=false. Elapsed: 18.37287857s
Jul 18 19:55:01.262: INFO: Pod "pod-db261340-640f-482a-8c8f-4ccd9a736f61": Phase="Pending", Reason="", readiness=false. Elapsed: 20.405226296s
Jul 18 19:55:03.294: INFO: Pod "pod-db261340-640f-482a-8c8f-4ccd9a736f61": Phase="Succeeded", Reason="", readiness=false. Elapsed: 22.437237297s
STEP: Saw pod success
Jul 18 19:55:03.294: INFO: Pod "pod-db261340-640f-482a-8c8f-4ccd9a736f61" satisfied condition "Succeeded or Failed"
Jul 18 19:55:03.326: INFO: Trying to get logs from node 10.116.0.5 pod pod-db261340-640f-482a-8c8f-4ccd9a736f61 container test-container: <nil>
STEP: delete the pod
Jul 18 19:55:03.408: INFO: Waiting for pod pod-db261340-640f-482a-8c8f-4ccd9a736f61 to disappear
Jul 18 19:55:03.438: INFO: Pod pod-db261340-640f-482a-8c8f-4ccd9a736f61 no longer exists
[AfterEach] [sig-storage] EmptyDir volumes
  /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:186
... skipping 4 lines ...
• [SLOW TEST:22.864 seconds]
[sig-storage] EmptyDir volumes
/workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/common/storage/framework.go:23
  should support (root,0666,tmpfs) [LinuxOnly] [NodeConformance] [Conformance]
  /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:630
------------------------------
{"msg":"PASSED [sig-storage] EmptyDir volumes should support (root,0666,tmpfs) [LinuxOnly] [NodeConformance] [Conformance]","total":-1,"completed":8,"skipped":36,"failed":0}

S
------------------------------
[BeforeEach] [Testpattern: Inline-volume (default fs)] volumes
  /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/storage/framework/testsuite.go:51
Jul 18 19:55:03.548: INFO: Only supported for providers [gce gke] (not skeleton)
... skipping 45 lines ...
STEP: Building a namespace api object, basename security-context-test
STEP: Waiting for a default service account to be provisioned in namespace
[BeforeEach] [sig-node] Security Context
  /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/common/node/security_context.go:46
[It] should allow privilege escalation when true [LinuxOnly] [NodeConformance]
  /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/common/node/security_context.go:367
Jul 18 19:54:47.607: INFO: Waiting up to 5m0s for pod "alpine-nnp-true-c20664be-f7fc-4f8c-80ad-b7e2fcbb0a75" in namespace "security-context-test-8878" to be "Succeeded or Failed"
Jul 18 19:54:47.638: INFO: Pod "alpine-nnp-true-c20664be-f7fc-4f8c-80ad-b7e2fcbb0a75": Phase="Pending", Reason="", readiness=false. Elapsed: 30.840863ms
Jul 18 19:54:49.673: INFO: Pod "alpine-nnp-true-c20664be-f7fc-4f8c-80ad-b7e2fcbb0a75": Phase="Pending", Reason="", readiness=false. Elapsed: 2.066165745s
Jul 18 19:54:51.706: INFO: Pod "alpine-nnp-true-c20664be-f7fc-4f8c-80ad-b7e2fcbb0a75": Phase="Pending", Reason="", readiness=false. Elapsed: 4.098853584s
Jul 18 19:54:53.739: INFO: Pod "alpine-nnp-true-c20664be-f7fc-4f8c-80ad-b7e2fcbb0a75": Phase="Pending", Reason="", readiness=false. Elapsed: 6.1316832s
Jul 18 19:54:55.772: INFO: Pod "alpine-nnp-true-c20664be-f7fc-4f8c-80ad-b7e2fcbb0a75": Phase="Pending", Reason="", readiness=false. Elapsed: 8.164440608s
Jul 18 19:54:57.807: INFO: Pod "alpine-nnp-true-c20664be-f7fc-4f8c-80ad-b7e2fcbb0a75": Phase="Pending", Reason="", readiness=false. Elapsed: 10.199916071s
Jul 18 19:54:59.838: INFO: Pod "alpine-nnp-true-c20664be-f7fc-4f8c-80ad-b7e2fcbb0a75": Phase="Pending", Reason="", readiness=false. Elapsed: 12.230574988s
Jul 18 19:55:01.875: INFO: Pod "alpine-nnp-true-c20664be-f7fc-4f8c-80ad-b7e2fcbb0a75": Phase="Pending", Reason="", readiness=false. Elapsed: 14.267818526s
Jul 18 19:55:03.907: INFO: Pod "alpine-nnp-true-c20664be-f7fc-4f8c-80ad-b7e2fcbb0a75": Phase="Succeeded", Reason="", readiness=false. Elapsed: 16.300143624s
Jul 18 19:55:03.907: INFO: Pod "alpine-nnp-true-c20664be-f7fc-4f8c-80ad-b7e2fcbb0a75" satisfied condition "Succeeded or Failed"
[AfterEach] [sig-node] Security Context
  /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:186
Jul 18 19:55:03.941: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready
STEP: Destroying namespace "security-context-test-8878" for this suite.


... skipping 2 lines ...
/workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/common/node/framework.go:23
  when creating containers with AllowPrivilegeEscalation
  /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/common/node/security_context.go:296
    should allow privilege escalation when true [LinuxOnly] [NodeConformance]
    /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/common/node/security_context.go:367
------------------------------
{"msg":"PASSED [sig-node] Security Context when creating containers with AllowPrivilegeEscalation should allow privilege escalation when true [LinuxOnly] [NodeConformance]","total":-1,"completed":11,"skipped":87,"failed":0}

SSSSSSSSSS
------------------------------
[BeforeEach] [Testpattern: Inline-volume (default fs)] subPath
  /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/storage/framework/testsuite.go:51
[BeforeEach] [Testpattern: Inline-volume (default fs)] subPath
... skipping 5 lines ...
[It] should support readOnly file specified in the volumeMount [LinuxOnly]
  /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/storage/testsuites/subpath.go:379
Jul 18 19:54:38.301: INFO: In-tree plugin kubernetes.io/empty-dir is not migrated, not validating any metrics
Jul 18 19:54:38.301: INFO: Creating resource for inline volume
STEP: Creating pod pod-subpath-test-inlinevolume-p2gk
STEP: Creating a pod to test subpath
Jul 18 19:54:38.337: INFO: Waiting up to 5m0s for pod "pod-subpath-test-inlinevolume-p2gk" in namespace "provisioning-1104" to be "Succeeded or Failed"
Jul 18 19:54:38.374: INFO: Pod "pod-subpath-test-inlinevolume-p2gk": Phase="Pending", Reason="", readiness=false. Elapsed: 36.234372ms
Jul 18 19:54:40.404: INFO: Pod "pod-subpath-test-inlinevolume-p2gk": Phase="Pending", Reason="", readiness=false. Elapsed: 2.066653373s
Jul 18 19:54:42.436: INFO: Pod "pod-subpath-test-inlinevolume-p2gk": Phase="Pending", Reason="", readiness=false. Elapsed: 4.098282029s
Jul 18 19:54:44.467: INFO: Pod "pod-subpath-test-inlinevolume-p2gk": Phase="Pending", Reason="", readiness=false. Elapsed: 6.129518958s
Jul 18 19:54:46.499: INFO: Pod "pod-subpath-test-inlinevolume-p2gk": Phase="Pending", Reason="", readiness=false. Elapsed: 8.161136865s
Jul 18 19:54:48.529: INFO: Pod "pod-subpath-test-inlinevolume-p2gk": Phase="Pending", Reason="", readiness=false. Elapsed: 10.191616885s
... skipping 3 lines ...
Jul 18 19:54:56.658: INFO: Pod "pod-subpath-test-inlinevolume-p2gk": Phase="Pending", Reason="", readiness=false. Elapsed: 18.320566615s
Jul 18 19:54:58.692: INFO: Pod "pod-subpath-test-inlinevolume-p2gk": Phase="Pending", Reason="", readiness=false. Elapsed: 20.354338313s
Jul 18 19:55:00.724: INFO: Pod "pod-subpath-test-inlinevolume-p2gk": Phase="Pending", Reason="", readiness=false. Elapsed: 22.38681543s
Jul 18 19:55:02.758: INFO: Pod "pod-subpath-test-inlinevolume-p2gk": Phase="Pending", Reason="", readiness=false. Elapsed: 24.420256374s
Jul 18 19:55:04.789: INFO: Pod "pod-subpath-test-inlinevolume-p2gk": Phase="Succeeded", Reason="", readiness=false. Elapsed: 26.451763675s
STEP: Saw pod success
Jul 18 19:55:04.789: INFO: Pod "pod-subpath-test-inlinevolume-p2gk" satisfied condition "Succeeded or Failed"
Jul 18 19:55:04.819: INFO: Trying to get logs from node 10.116.0.5 pod pod-subpath-test-inlinevolume-p2gk container test-container-subpath-inlinevolume-p2gk: <nil>
STEP: delete the pod
Jul 18 19:55:04.895: INFO: Waiting for pod pod-subpath-test-inlinevolume-p2gk to disappear
Jul 18 19:55:04.924: INFO: Pod pod-subpath-test-inlinevolume-p2gk no longer exists
STEP: Deleting pod pod-subpath-test-inlinevolume-p2gk
Jul 18 19:55:04.925: INFO: Deleting pod "pod-subpath-test-inlinevolume-p2gk" in namespace "provisioning-1104"
... skipping 12 lines ...
  /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/storage/in_tree_volumes.go:58
    [Testpattern: Inline-volume (default fs)] subPath
    /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/storage/framework/testsuite.go:50
      should support readOnly file specified in the volumeMount [LinuxOnly]
      /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/storage/testsuites/subpath.go:379
------------------------------
{"msg":"PASSED [sig-storage] In-tree Volumes [Driver: emptydir] [Testpattern: Inline-volume (default fs)] subPath should support readOnly file specified in the volumeMount [LinuxOnly]","total":-1,"completed":8,"skipped":44,"failed":0}

S
------------------------------
[BeforeEach] [Testpattern: Dynamic PV (default fs)] subPath
  /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/storage/framework/testsuite.go:51
Jul 18 19:55:05.097: INFO: Driver hostPathSymlink doesn't support DynamicPV -- skipping
... skipping 60 lines ...
STEP: Destroying namespace "services-8703" for this suite.
[AfterEach] [sig-network] Services
  /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/network/service.go:750

•
------------------------------
{"msg":"PASSED [sig-network] Services should check NodePort out-of-range","total":-1,"completed":9,"skipped":58,"failed":0}

S
------------------------------
[BeforeEach] [Testpattern: Inline-volume (ext4)] volumes
  /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/storage/framework/testsuite.go:51
Jul 18 19:55:05.604: INFO: Only supported for providers [gce gke] (not skeleton)
... skipping 43 lines ...
STEP: Building a namespace api object, basename configmap
STEP: Waiting for a default service account to be provisioned in namespace
[It] should be consumable from pods in volume with mappings and Item mode set [LinuxOnly] [NodeConformance] [Conformance]
  /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:630
STEP: Creating configMap with name configmap-test-volume-map-1634e87f-ea7a-4c59-8067-bb717b2d1b38
STEP: Creating a pod to test consume configMaps
Jul 18 19:54:59.807: INFO: Waiting up to 5m0s for pod "pod-configmaps-841dd32f-d005-410e-947a-6c2d9f485640" in namespace "configmap-5568" to be "Succeeded or Failed"
Jul 18 19:54:59.837: INFO: Pod "pod-configmaps-841dd32f-d005-410e-947a-6c2d9f485640": Phase="Pending", Reason="", readiness=false. Elapsed: 30.312278ms
Jul 18 19:55:01.869: INFO: Pod "pod-configmaps-841dd32f-d005-410e-947a-6c2d9f485640": Phase="Pending", Reason="", readiness=false. Elapsed: 2.062341637s
Jul 18 19:55:03.901: INFO: Pod "pod-configmaps-841dd32f-d005-410e-947a-6c2d9f485640": Phase="Pending", Reason="", readiness=false. Elapsed: 4.093937204s
Jul 18 19:55:05.932: INFO: Pod "pod-configmaps-841dd32f-d005-410e-947a-6c2d9f485640": Phase="Pending", Reason="", readiness=false. Elapsed: 6.124999197s
Jul 18 19:55:07.963: INFO: Pod "pod-configmaps-841dd32f-d005-410e-947a-6c2d9f485640": Phase="Pending", Reason="", readiness=false. Elapsed: 8.156528835s
Jul 18 19:55:09.995: INFO: Pod "pod-configmaps-841dd32f-d005-410e-947a-6c2d9f485640": Phase="Succeeded", Reason="", readiness=false. Elapsed: 10.188190319s
STEP: Saw pod success
Jul 18 19:55:09.995: INFO: Pod "pod-configmaps-841dd32f-d005-410e-947a-6c2d9f485640" satisfied condition "Succeeded or Failed"
Jul 18 19:55:10.025: INFO: Trying to get logs from node 10.116.0.5 pod pod-configmaps-841dd32f-d005-410e-947a-6c2d9f485640 container agnhost-container: <nil>
STEP: delete the pod
Jul 18 19:55:10.100: INFO: Waiting for pod pod-configmaps-841dd32f-d005-410e-947a-6c2d9f485640 to disappear
Jul 18 19:55:10.344: INFO: Pod pod-configmaps-841dd32f-d005-410e-947a-6c2d9f485640 no longer exists
[AfterEach] [sig-storage] ConfigMap
  /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:186
... skipping 4 lines ...
• [SLOW TEST:10.857 seconds]
[sig-storage] ConfigMap
/workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/common/storage/framework.go:23
  should be consumable from pods in volume with mappings and Item mode set [LinuxOnly] [NodeConformance] [Conformance]
  /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:630
------------------------------
{"msg":"PASSED [sig-storage] ConfigMap should be consumable from pods in volume with mappings and Item mode set [LinuxOnly] [NodeConformance] [Conformance]","total":-1,"completed":8,"skipped":71,"failed":0}

S
------------------------------
[BeforeEach] [sig-cli] Kubectl Port forwarding
  /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:185
STEP: Creating a kubernetes client
... skipping 47 lines ...
  /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/kubectl/portforward.go:452
    that expects a client request
    /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/kubectl/portforward.go:453
      should support a client that connects, sends NO DATA, and disconnects
      /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/kubectl/portforward.go:454
------------------------------
{"msg":"PASSED [sig-cli] Kubectl Port forwarding With a server listening on 0.0.0.0 that expects a client request should support a client that connects, sends NO DATA, and disconnects","total":-1,"completed":9,"skipped":88,"failed":0}

S
------------------------------
[BeforeEach] [sig-api-machinery] Servers with support for Table transformation
  /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:185
STEP: Creating a kubernetes client
... skipping 17 lines ...
  /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:186
Jul 18 19:55:12.089: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready
STEP: Destroying namespace "tables-1096" for this suite.

•
------------------------------
{"msg":"PASSED [sig-api-machinery] Servers with support for Table transformation should return generic metadata details across all namespaces for nodes","total":-1,"completed":10,"skipped":89,"failed":0}

SSSSSSSS
------------------------------
[BeforeEach] [Testpattern: Pre-provisioned PV (block volmode)] volumeMode
  /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/storage/framework/testsuite.go:51
[BeforeEach] [Testpattern: Pre-provisioned PV (block volmode)] volumeMode
... skipping 92 lines ...
  /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/storage/persistent_volumes-local.go:192
    Two pods mounting a local volume one after the other
    /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/storage/persistent_volumes-local.go:254
      should be able to write from pod1 and read from pod2
      /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/storage/persistent_volumes-local.go:255
------------------------------
{"msg":"PASSED [sig-storage] PersistentVolumes-local  [Volume type: dir-link] Two pods mounting a local volume one after the other should be able to write from pod1 and read from pod2","total":-1,"completed":11,"skipped":34,"failed":0}

SS
------------------------------
[BeforeEach] [sig-storage] Projected downwardAPI
  /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:185
STEP: Creating a kubernetes client
... skipping 2 lines ...
STEP: Waiting for a default service account to be provisioned in namespace
[BeforeEach] [sig-storage] Projected downwardAPI
  /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/common/storage/projected_downwardapi.go:41
[It] should provide container's memory request [NodeConformance] [Conformance]
  /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:630
STEP: Creating a pod to test downward API volume plugin
Jul 18 19:55:03.757: INFO: Waiting up to 5m0s for pod "downwardapi-volume-3835bed3-916b-41b9-8083-1f0da241d8fb" in namespace "projected-4675" to be "Succeeded or Failed"
Jul 18 19:55:03.788: INFO: Pod "downwardapi-volume-3835bed3-916b-41b9-8083-1f0da241d8fb": Phase="Pending", Reason="", readiness=false. Elapsed: 30.650406ms
Jul 18 19:55:05.819: INFO: Pod "downwardapi-volume-3835bed3-916b-41b9-8083-1f0da241d8fb": Phase="Pending", Reason="", readiness=false. Elapsed: 2.062046708s
Jul 18 19:55:07.874: INFO: Pod "downwardapi-volume-3835bed3-916b-41b9-8083-1f0da241d8fb": Phase="Pending", Reason="", readiness=false. Elapsed: 4.116761467s
Jul 18 19:55:09.906: INFO: Pod "downwardapi-volume-3835bed3-916b-41b9-8083-1f0da241d8fb": Phase="Pending", Reason="", readiness=false. Elapsed: 6.14912915s
Jul 18 19:55:11.940: INFO: Pod "downwardapi-volume-3835bed3-916b-41b9-8083-1f0da241d8fb": Phase="Pending", Reason="", readiness=false. Elapsed: 8.182181716s
Jul 18 19:55:13.972: INFO: Pod "downwardapi-volume-3835bed3-916b-41b9-8083-1f0da241d8fb": Phase="Succeeded", Reason="", readiness=false. Elapsed: 10.214219719s
STEP: Saw pod success
Jul 18 19:55:13.972: INFO: Pod "downwardapi-volume-3835bed3-916b-41b9-8083-1f0da241d8fb" satisfied condition "Succeeded or Failed"
Jul 18 19:55:14.002: INFO: Trying to get logs from node 10.116.0.5 pod downwardapi-volume-3835bed3-916b-41b9-8083-1f0da241d8fb container client-container: <nil>
STEP: delete the pod
Jul 18 19:55:14.082: INFO: Waiting for pod downwardapi-volume-3835bed3-916b-41b9-8083-1f0da241d8fb to disappear
Jul 18 19:55:14.112: INFO: Pod downwardapi-volume-3835bed3-916b-41b9-8083-1f0da241d8fb no longer exists
[AfterEach] [sig-storage] Projected downwardAPI
  /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:186
... skipping 4 lines ...
• [SLOW TEST:10.639 seconds]
[sig-storage] Projected downwardAPI
/workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/common/storage/framework.go:23
  should provide container's memory request [NodeConformance] [Conformance]
  /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:630
------------------------------
{"msg":"PASSED [sig-storage] Projected downwardAPI should provide container's memory request [NodeConformance] [Conformance]","total":-1,"completed":9,"skipped":40,"failed":0}

S
------------------------------
[BeforeEach] [Testpattern: Pre-provisioned PV (default fs)] volumes
  /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/storage/framework/testsuite.go:51
[BeforeEach] [Testpattern: Pre-provisioned PV (default fs)] volumes
... skipping 16 lines ...
Jul 18 19:55:06.437: INFO: PersistentVolumeClaim pvc-8q6sl found but phase is Pending instead of Bound.
Jul 18 19:55:08.470: INFO: PersistentVolumeClaim pvc-8q6sl found and phase=Bound (4.096879174s)
Jul 18 19:55:08.470: INFO: Waiting up to 3m0s for PersistentVolume local-484cg to have phase Bound
Jul 18 19:55:08.501: INFO: PersistentVolume local-484cg found and phase=Bound (31.405506ms)
STEP: Creating pod exec-volume-test-preprovisionedpv-ddtw
STEP: Creating a pod to test exec-volume-test
Jul 18 19:55:08.607: INFO: Waiting up to 5m0s for pod "exec-volume-test-preprovisionedpv-ddtw" in namespace "volume-2322" to be "Succeeded or Failed"
Jul 18 19:55:08.639: INFO: Pod "exec-volume-test-preprovisionedpv-ddtw": Phase="Pending", Reason="", readiness=false. Elapsed: 32.02083ms
Jul 18 19:55:10.672: INFO: Pod "exec-volume-test-preprovisionedpv-ddtw": Phase="Pending", Reason="", readiness=false. Elapsed: 2.065133163s
Jul 18 19:55:12.705: INFO: Pod "exec-volume-test-preprovisionedpv-ddtw": Phase="Pending", Reason="", readiness=false. Elapsed: 4.098557097s
Jul 18 19:55:14.738: INFO: Pod "exec-volume-test-preprovisionedpv-ddtw": Phase="Succeeded", Reason="", readiness=false. Elapsed: 6.130790071s
STEP: Saw pod success
Jul 18 19:55:14.738: INFO: Pod "exec-volume-test-preprovisionedpv-ddtw" satisfied condition "Succeeded or Failed"
Jul 18 19:55:14.769: INFO: Trying to get logs from node 10.116.0.5 pod exec-volume-test-preprovisionedpv-ddtw container exec-container-preprovisionedpv-ddtw: <nil>
STEP: delete the pod
Jul 18 19:55:14.851: INFO: Waiting for pod exec-volume-test-preprovisionedpv-ddtw to disappear
Jul 18 19:55:14.884: INFO: Pod exec-volume-test-preprovisionedpv-ddtw no longer exists
STEP: Deleting pod exec-volume-test-preprovisionedpv-ddtw
Jul 18 19:55:14.884: INFO: Deleting pod "exec-volume-test-preprovisionedpv-ddtw" in namespace "volume-2322"
... skipping 17 lines ...
  /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/storage/in_tree_volumes.go:58
    [Testpattern: Pre-provisioned PV (default fs)] volumes
    /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/storage/framework/testsuite.go:50
      should allow exec of files on the volume
      /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/storage/testsuites/volumes.go:196
------------------------------
{"msg":"PASSED [sig-storage] In-tree Volumes [Driver: local][LocalVolumeType: dir] [Testpattern: Pre-provisioned PV (default fs)] volumes should allow exec of files on the volume","total":-1,"completed":4,"skipped":21,"failed":0}

SS
------------------------------
[BeforeEach] [Testpattern: Inline-volume (ext4)] volumes
  /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/storage/framework/testsuite.go:51
Jul 18 19:55:15.386: INFO: Driver local doesn't support InlineVolume -- skipping
... skipping 56 lines ...
• [SLOW TEST:10.342 seconds]
[sig-network] Services
/workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/network/common/framework.go:23
  should allow pods to hairpin back to themselves through services
  /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/network/service.go:986
------------------------------
{"msg":"PASSED [sig-network] Services should allow pods to hairpin back to themselves through services","total":-1,"completed":10,"skipped":61,"failed":0}

SS
------------------------------
[BeforeEach] [Testpattern: Dynamic PV (default fs)] capacity
  /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/storage/framework/testsuite.go:51
Jul 18 19:55:15.983: INFO: Only supported for providers [azure] (not skeleton)
... skipping 167 lines ...
Jul 18 19:55:10.452: INFO: >>> kubeConfig: /root/.kube/config
STEP: Building a namespace api object, basename downward-api
STEP: Waiting for a default service account to be provisioned in namespace
[It] should provide default limits.cpu/memory from node allocatable [NodeConformance] [Conformance]
  /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:630
STEP: Creating a pod to test downward api env vars
Jul 18 19:55:10.650: INFO: Waiting up to 5m0s for pod "downward-api-8b4598c4-8ce7-4da6-8f7a-89cfec635f76" in namespace "downward-api-9618" to be "Succeeded or Failed"
Jul 18 19:55:10.680: INFO: Pod "downward-api-8b4598c4-8ce7-4da6-8f7a-89cfec635f76": Phase="Pending", Reason="", readiness=false. Elapsed: 30.571542ms
Jul 18 19:55:12.712: INFO: Pod "downward-api-8b4598c4-8ce7-4da6-8f7a-89cfec635f76": Phase="Pending", Reason="", readiness=false. Elapsed: 2.061865261s
Jul 18 19:55:14.743: INFO: Pod "downward-api-8b4598c4-8ce7-4da6-8f7a-89cfec635f76": Phase="Pending", Reason="", readiness=false. Elapsed: 4.093078283s
Jul 18 19:55:16.777: INFO: Pod "downward-api-8b4598c4-8ce7-4da6-8f7a-89cfec635f76": Phase="Succeeded", Reason="", readiness=false. Elapsed: 6.126848297s
STEP: Saw pod success
Jul 18 19:55:16.777: INFO: Pod "downward-api-8b4598c4-8ce7-4da6-8f7a-89cfec635f76" satisfied condition "Succeeded or Failed"
Jul 18 19:55:16.807: INFO: Trying to get logs from node 10.116.0.5 pod downward-api-8b4598c4-8ce7-4da6-8f7a-89cfec635f76 container dapi-container: <nil>
STEP: delete the pod
Jul 18 19:55:16.890: INFO: Waiting for pod downward-api-8b4598c4-8ce7-4da6-8f7a-89cfec635f76 to disappear
Jul 18 19:55:16.921: INFO: Pod downward-api-8b4598c4-8ce7-4da6-8f7a-89cfec635f76 no longer exists
[AfterEach] [sig-node] Downward API
  /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:186
... skipping 4 lines ...
• [SLOW TEST:6.561 seconds]
[sig-node] Downward API
/workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/common/node/framework.go:23
  should provide default limits.cpu/memory from node allocatable [NodeConformance] [Conformance]
  /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:630
------------------------------
{"msg":"PASSED [sig-node] Downward API should provide default limits.cpu/memory from node allocatable [NodeConformance] [Conformance]","total":-1,"completed":9,"skipped":72,"failed":0}

SSSSSSSSSSSSS
------------------------------
[BeforeEach] [Testpattern: Dynamic PV (default fs)(allowExpansion)] volume-expand
  /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/storage/framework/testsuite.go:51
Jul 18 19:55:17.070: INFO: Driver local doesn't support DynamicPV -- skipping
... skipping 132 lines ...
/workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/storage/utils/framework.go:23
  CSI FSGroupPolicy [LinuxOnly]
  /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/storage/csi_mock_volume.go:1436
    should modify fsGroup if fsGroupPolicy=default
    /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/storage/csi_mock_volume.go:1460
------------------------------
{"msg":"PASSED [sig-storage] CSI mock volume CSI FSGroupPolicy [LinuxOnly] should modify fsGroup if fsGroupPolicy=default","total":-1,"completed":4,"skipped":53,"failed":0}

SSSS
------------------------------
{"msg":"PASSED [sig-storage] CSI mock volume CSIStorageCapacity CSIStorageCapacity used, have capacity","total":-1,"completed":9,"skipped":37,"failed":0}
[BeforeEach] [Testpattern: Pre-provisioned PV (default fs)] subPath
  /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/storage/framework/testsuite.go:51
[BeforeEach] [Testpattern: Pre-provisioned PV (default fs)] subPath
  /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:185
STEP: Creating a kubernetes client
Jul 18 19:54:56.192: INFO: >>> kubeConfig: /root/.kube/config
... skipping 17 lines ...
Jul 18 19:55:05.722: INFO: PersistentVolumeClaim pvc-92f9m found but phase is Pending instead of Bound.
Jul 18 19:55:07.753: INFO: PersistentVolumeClaim pvc-92f9m found and phase=Bound (2.060877205s)
Jul 18 19:55:07.753: INFO: Waiting up to 3m0s for PersistentVolume local-8v77k to have phase Bound
Jul 18 19:55:07.784: INFO: PersistentVolume local-8v77k found and phase=Bound (30.799285ms)
STEP: Creating pod pod-subpath-test-preprovisionedpv-cjrg
STEP: Creating a pod to test subpath
Jul 18 19:55:07.879: INFO: Waiting up to 5m0s for pod "pod-subpath-test-preprovisionedpv-cjrg" in namespace "provisioning-1241" to be "Succeeded or Failed"
Jul 18 19:55:07.910: INFO: Pod "pod-subpath-test-preprovisionedpv-cjrg": Phase="Pending", Reason="", readiness=false. Elapsed: 30.235292ms
Jul 18 19:55:09.945: INFO: Pod "pod-subpath-test-preprovisionedpv-cjrg": Phase="Pending", Reason="", readiness=false. Elapsed: 2.065669471s
Jul 18 19:55:11.976: INFO: Pod "pod-subpath-test-preprovisionedpv-cjrg": Phase="Pending", Reason="", readiness=false. Elapsed: 4.096692014s
Jul 18 19:55:14.007: INFO: Pod "pod-subpath-test-preprovisionedpv-cjrg": Phase="Pending", Reason="", readiness=false. Elapsed: 6.127578908s
Jul 18 19:55:16.040: INFO: Pod "pod-subpath-test-preprovisionedpv-cjrg": Phase="Pending", Reason="", readiness=false. Elapsed: 8.160574057s
Jul 18 19:55:18.071: INFO: Pod "pod-subpath-test-preprovisionedpv-cjrg": Phase="Pending", Reason="", readiness=false. Elapsed: 10.192020328s
Jul 18 19:55:20.103: INFO: Pod "pod-subpath-test-preprovisionedpv-cjrg": Phase="Succeeded", Reason="", readiness=false. Elapsed: 12.223910483s
STEP: Saw pod success
Jul 18 19:55:20.103: INFO: Pod "pod-subpath-test-preprovisionedpv-cjrg" satisfied condition "Succeeded or Failed"
Jul 18 19:55:20.134: INFO: Trying to get logs from node 10.116.0.7 pod pod-subpath-test-preprovisionedpv-cjrg container test-container-volume-preprovisionedpv-cjrg: <nil>
STEP: delete the pod
Jul 18 19:55:20.216: INFO: Waiting for pod pod-subpath-test-preprovisionedpv-cjrg to disappear
Jul 18 19:55:20.245: INFO: Pod pod-subpath-test-preprovisionedpv-cjrg no longer exists
STEP: Deleting pod pod-subpath-test-preprovisionedpv-cjrg
Jul 18 19:55:20.245: INFO: Deleting pod "pod-subpath-test-preprovisionedpv-cjrg" in namespace "provisioning-1241"
... skipping 28 lines ...
    /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/storage/framework/testsuite.go:50
      should support existing directory
      /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/storage/testsuites/subpath.go:205
------------------------------
SS
------------------------------
{"msg":"PASSED [sig-storage] In-tree Volumes [Driver: local][LocalVolumeType: blockfs] [Testpattern: Pre-provisioned PV (default fs)] subPath should support existing directory","total":-1,"completed":10,"skipped":37,"failed":0}

SS
------------------------------
[BeforeEach] [Testpattern: Dynamic PV (delayed binding)] topology
  /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/storage/framework/testsuite.go:51
Jul 18 19:55:21.570: INFO: Only supported for providers [gce gke] (not skeleton)
... skipping 5 lines ...
[sig-storage] In-tree Volumes
/workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/storage/utils/framework.go:23
  [Driver: windows-gcepd]
  /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/storage/in_tree_volumes.go:58
    [Testpattern: Dynamic PV (delayed binding)] topology
    /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/storage/framework/testsuite.go:50
      should fail to schedule a pod which has topologies that conflict with AllowedTopologies [BeforeEach]
      /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/storage/testsuites/topology.go:192

      Only supported for providers [gce gke] (not skeleton)

      /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/storage/drivers/in_tree.go:1301
------------------------------
... skipping 90 lines ...
STEP: Destroying namespace "services-3593" for this suite.
[AfterEach] [sig-network] Services
  /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/network/service.go:750

•
------------------------------
{"msg":"PASSED [sig-network] Services should complete a service status lifecycle [Conformance]","total":-1,"completed":11,"skipped":39,"failed":0}
[BeforeEach] [Testpattern: Dynamic PV (delayed binding)] topology
  /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/storage/framework/testsuite.go:51
Jul 18 19:55:22.478: INFO: Only supported for providers [gce gke] (not skeleton)
[AfterEach] [Testpattern: Dynamic PV (delayed binding)] topology
  /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:186

... skipping 18 lines ...
Jul 18 19:54:48.100: INFO: >>> kubeConfig: /root/.kube/config
STEP: Building a namespace api object, basename svcaccounts
STEP: Waiting for a default service account to be provisioned in namespace
[It] ServiceAccountIssuerDiscovery should support OIDC discovery of service account issuer [Conformance]
  /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:630
Jul 18 19:54:48.328: INFO: created pod
Jul 18 19:54:48.328: INFO: Waiting up to 5m0s for pod "oidc-discovery-validator" in namespace "svcaccounts-4247" to be "Succeeded or Failed"
Jul 18 19:54:48.361: INFO: Pod "oidc-discovery-validator": Phase="Pending", Reason="", readiness=false. Elapsed: 32.855517ms
Jul 18 19:54:50.393: INFO: Pod "oidc-discovery-validator": Phase="Pending", Reason="", readiness=false. Elapsed: 2.064472367s
Jul 18 19:54:52.427: INFO: Pod "oidc-discovery-validator": Phase="Pending", Reason="", readiness=false. Elapsed: 4.098742444s
Jul 18 19:54:54.460: INFO: Pod "oidc-discovery-validator": Phase="Pending", Reason="", readiness=false. Elapsed: 6.131536841s
Jul 18 19:54:56.492: INFO: Pod "oidc-discovery-validator": Phase="Pending", Reason="", readiness=false. Elapsed: 8.163704837s
Jul 18 19:54:58.523: INFO: Pod "oidc-discovery-validator": Phase="Succeeded", Reason="", readiness=false. Elapsed: 10.195176078s
STEP: Saw pod success
Jul 18 19:54:58.524: INFO: Pod "oidc-discovery-validator" satisfied condition "Succeeded or Failed"
Jul 18 19:55:28.524: INFO: polling logs
Jul 18 19:55:28.557: INFO: Pod logs: 
2021/07/18 19:54:49 OK: Got token
2021/07/18 19:54:49 validating with in-cluster discovery
2021/07/18 19:54:49 OK: got issuer https://kubernetes.default
2021/07/18 19:54:49 Full, not-validated claims: 
... skipping 14 lines ...
• [SLOW TEST:40.586 seconds]
[sig-auth] ServiceAccounts
/workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/auth/framework.go:23
  ServiceAccountIssuerDiscovery should support OIDC discovery of service account issuer [Conformance]
  /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:630
------------------------------
{"msg":"PASSED [sig-auth] ServiceAccounts ServiceAccountIssuerDiscovery should support OIDC discovery of service account issuer [Conformance]","total":-1,"completed":13,"skipped":76,"failed":0}
[BeforeEach] [Testpattern: Inline-volume (default fs)] volumes
  /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/storage/framework/testsuite.go:51
Jul 18 19:55:28.701: INFO: Driver local doesn't support InlineVolume -- skipping
[AfterEach] [Testpattern: Inline-volume (default fs)] volumes
  /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:186

... skipping 151 lines ...
• [SLOW TEST:11.851 seconds]
[sig-api-machinery] ResourceQuota
/workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/apimachinery/framework.go:23
  should create a ResourceQuota and capture the life of a persistent volume claim with a storage class
  /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/apimachinery/resource_quota.go:531
------------------------------
{"msg":"PASSED [sig-api-machinery] ResourceQuota should create a ResourceQuota and capture the life of a persistent volume claim with a storage class","total":-1,"completed":10,"skipped":90,"failed":0}

SSS
------------------------------
[BeforeEach] [sig-api-machinery] AdmissionWebhook [Privileged:ClusterAdmin]
  /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:185
STEP: Creating a kubernetes client
... skipping 29 lines ...
• [SLOW TEST:8.772 seconds]
[sig-api-machinery] AdmissionWebhook [Privileged:ClusterAdmin]
/workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/apimachinery/framework.go:23
  should mutate configmap [Conformance]
  /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:630
------------------------------
{"msg":"PASSED [sig-api-machinery] AdmissionWebhook [Privileged:ClusterAdmin] should mutate configmap [Conformance]","total":-1,"completed":5,"skipped":66,"failed":0}

S
------------------------------
[BeforeEach] [Testpattern: Pre-provisioned PV (default fs)] subPath
  /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/storage/framework/testsuite.go:51
Jul 18 19:55:30.392: INFO: Driver hostPath doesn't support PreprovisionedPV -- skipping
... skipping 38 lines ...
• [SLOW TEST:18.309 seconds]
[sig-api-machinery] Garbage collector
/workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/apimachinery/framework.go:23
  should support cascading deletion of custom resources
  /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/apimachinery/garbage_collector.go:920
------------------------------
{"msg":"PASSED [sig-api-machinery] Garbage collector should support cascading deletion of custom resources","total":-1,"completed":11,"skipped":99,"failed":0}

SSSS
------------------------------
[BeforeEach] [Testpattern: Inline-volume (default fs)] volumes
  /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/storage/framework/testsuite.go:51
Jul 18 19:55:30.790: INFO: Only supported for providers [vsphere] (not skeleton)
... skipping 116 lines ...
STEP: Building a namespace api object, basename projected
STEP: Waiting for a default service account to be provisioned in namespace
[It] should be consumable from pods in volume with mappings [NodeConformance] [Conformance]
  /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:630
STEP: Creating projection with secret that has name projected-secret-test-map-6224d14e-7483-494d-b7ad-003b6dcc4e8e
STEP: Creating a pod to test consume secrets
Jul 18 19:55:22.792: INFO: Waiting up to 5m0s for pod "pod-projected-secrets-c63f90ff-3b3e-4e52-b79d-ed23e6aba0e2" in namespace "projected-7802" to be "Succeeded or Failed"
Jul 18 19:55:22.830: INFO: Pod "pod-projected-secrets-c63f90ff-3b3e-4e52-b79d-ed23e6aba0e2": Phase="Pending", Reason="", readiness=false. Elapsed: 37.743428ms
Jul 18 19:55:24.861: INFO: Pod "pod-projected-secrets-c63f90ff-3b3e-4e52-b79d-ed23e6aba0e2": Phase="Pending", Reason="", readiness=false. Elapsed: 2.06826245s
Jul 18 19:55:26.891: INFO: Pod "pod-projected-secrets-c63f90ff-3b3e-4e52-b79d-ed23e6aba0e2": Phase="Pending", Reason="", readiness=false. Elapsed: 4.098945376s
Jul 18 19:55:28.923: INFO: Pod "pod-projected-secrets-c63f90ff-3b3e-4e52-b79d-ed23e6aba0e2": Phase="Pending", Reason="", readiness=false. Elapsed: 6.130461031s
Jul 18 19:55:30.957: INFO: Pod "pod-projected-secrets-c63f90ff-3b3e-4e52-b79d-ed23e6aba0e2": Phase="Pending", Reason="", readiness=false. Elapsed: 8.164840945s
Jul 18 19:55:32.995: INFO: Pod "pod-projected-secrets-c63f90ff-3b3e-4e52-b79d-ed23e6aba0e2": Phase="Succeeded", Reason="", readiness=false. Elapsed: 10.202743651s
STEP: Saw pod success
Jul 18 19:55:32.995: INFO: Pod "pod-projected-secrets-c63f90ff-3b3e-4e52-b79d-ed23e6aba0e2" satisfied condition "Succeeded or Failed"
Jul 18 19:55:33.025: INFO: Trying to get logs from node 10.116.0.5 pod pod-projected-secrets-c63f90ff-3b3e-4e52-b79d-ed23e6aba0e2 container projected-secret-volume-test: <nil>
STEP: delete the pod
Jul 18 19:55:33.100: INFO: Waiting for pod pod-projected-secrets-c63f90ff-3b3e-4e52-b79d-ed23e6aba0e2 to disappear
Jul 18 19:55:33.130: INFO: Pod pod-projected-secrets-c63f90ff-3b3e-4e52-b79d-ed23e6aba0e2 no longer exists
[AfterEach] [sig-storage] Projected secret
  /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:186
... skipping 4 lines ...
• [SLOW TEST:10.735 seconds]
[sig-storage] Projected secret
/workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/common/storage/framework.go:23
  should be consumable from pods in volume with mappings [NodeConformance] [Conformance]
  /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:630
------------------------------
{"msg":"PASSED [sig-storage] Projected secret should be consumable from pods in volume with mappings [NodeConformance] [Conformance]","total":-1,"completed":12,"skipped":40,"failed":0}

SS
------------------------------
[BeforeEach] [Testpattern: Dynamic PV (block volmode)] provisioning
  /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/storage/framework/testsuite.go:51
Jul 18 19:55:33.238: INFO: Driver hostPath doesn't support DynamicPV -- skipping
... skipping 222 lines ...
Jul 18 19:55:20.111: INFO: PersistentVolumeClaim pvc-knl2k found but phase is Pending instead of Bound.
Jul 18 19:55:22.143: INFO: PersistentVolumeClaim pvc-knl2k found and phase=Bound (14.263775512s)
Jul 18 19:55:22.143: INFO: Waiting up to 3m0s for PersistentVolume local-c5m5q to have phase Bound
Jul 18 19:55:22.173: INFO: PersistentVolume local-c5m5q found and phase=Bound (30.018527ms)
STEP: Creating pod pod-subpath-test-preprovisionedpv-hgjv
STEP: Creating a pod to test subpath
Jul 18 19:55:22.702: INFO: Waiting up to 5m0s for pod "pod-subpath-test-preprovisionedpv-hgjv" in namespace "provisioning-1486" to be "Succeeded or Failed"
Jul 18 19:55:22.733: INFO: Pod "pod-subpath-test-preprovisionedpv-hgjv": Phase="Pending", Reason="", readiness=false. Elapsed: 30.416761ms
Jul 18 19:55:24.764: INFO: Pod "pod-subpath-test-preprovisionedpv-hgjv": Phase="Pending", Reason="", readiness=false. Elapsed: 2.062009794s
Jul 18 19:55:26.795: INFO: Pod "pod-subpath-test-preprovisionedpv-hgjv": Phase="Pending", Reason="", readiness=false. Elapsed: 4.092952905s
Jul 18 19:55:28.827: INFO: Pod "pod-subpath-test-preprovisionedpv-hgjv": Phase="Pending", Reason="", readiness=false. Elapsed: 6.124653896s
Jul 18 19:55:30.861: INFO: Pod "pod-subpath-test-preprovisionedpv-hgjv": Phase="Pending", Reason="", readiness=false. Elapsed: 8.158724158s
Jul 18 19:55:32.894: INFO: Pod "pod-subpath-test-preprovisionedpv-hgjv": Phase="Pending", Reason="", readiness=false. Elapsed: 10.191259332s
Jul 18 19:55:34.925: INFO: Pod "pod-subpath-test-preprovisionedpv-hgjv": Phase="Succeeded", Reason="", readiness=false. Elapsed: 12.222226848s
STEP: Saw pod success
Jul 18 19:55:34.925: INFO: Pod "pod-subpath-test-preprovisionedpv-hgjv" satisfied condition "Succeeded or Failed"
Jul 18 19:55:34.957: INFO: Trying to get logs from node 10.116.0.5 pod pod-subpath-test-preprovisionedpv-hgjv container test-container-volume-preprovisionedpv-hgjv: <nil>
STEP: delete the pod
Jul 18 19:55:35.034: INFO: Waiting for pod pod-subpath-test-preprovisionedpv-hgjv to disappear
Jul 18 19:55:35.064: INFO: Pod pod-subpath-test-preprovisionedpv-hgjv no longer exists
STEP: Deleting pod pod-subpath-test-preprovisionedpv-hgjv
Jul 18 19:55:35.064: INFO: Deleting pod "pod-subpath-test-preprovisionedpv-hgjv" in namespace "provisioning-1486"
... skipping 19 lines ...
  /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/storage/in_tree_volumes.go:58
    [Testpattern: Pre-provisioned PV (default fs)] subPath
    /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/storage/framework/testsuite.go:50
      should support non-existent path
      /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/storage/testsuites/subpath.go:194
------------------------------
{"msg":"PASSED [sig-storage] In-tree Volumes [Driver: local][LocalVolumeType: dir] [Testpattern: Pre-provisioned PV (default fs)] subPath should support non-existent path","total":-1,"completed":10,"skipped":76,"failed":0}
[BeforeEach] [Testpattern: Inline-volume (default fs)] subPath
  /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/storage/framework/testsuite.go:51
Jul 18 19:55:35.599: INFO: Driver local doesn't support InlineVolume -- skipping
[AfterEach] [Testpattern: Inline-volume (default fs)] subPath
  /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:186

... skipping 114 lines ...
  /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/storage/persistent_volumes-local.go:192
    One pod requesting one prebound PVC
    /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/storage/persistent_volumes-local.go:209
      should be able to mount volume and read from pod1
      /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/storage/persistent_volumes-local.go:232
------------------------------
{"msg":"PASSED [sig-storage] PersistentVolumes-local  [Volume type: dir-bindmounted] One pod requesting one prebound PVC should be able to mount volume and read from pod1","total":-1,"completed":11,"skipped":90,"failed":0}

SSSSSSSSS
------------------------------
[BeforeEach] [Testpattern: Pre-provisioned PV (default fs)] subPath
  /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/storage/framework/testsuite.go:51
Jul 18 19:55:36.324: INFO: Only supported for providers [openstack] (not skeleton)
... skipping 14 lines ...
      Only supported for providers [openstack] (not skeleton)

      /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/storage/drivers/in_tree.go:1092
------------------------------
SSS
------------------------------
{"msg":"PASSED [sig-api-machinery] AdmissionWebhook [Privileged:ClusterAdmin] should mutate custom resource with different stored version [Conformance]","total":-1,"completed":12,"skipped":36,"failed":0}
[BeforeEach] [sig-api-machinery] CustomResourceDefinition resources [Privileged:ClusterAdmin]
  /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:185
STEP: Creating a kubernetes client
Jul 18 19:55:31.350: INFO: >>> kubeConfig: /root/.kube/config
STEP: Building a namespace api object, basename custom-resource-definition
STEP: Waiting for a default service account to be provisioned in namespace
... skipping 11 lines ...
/workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/apimachinery/framework.go:23
  Simple CustomResourceDefinition
  /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/apimachinery/custom_resource_definition.go:48
    getting/updating/patching custom resource definition status sub-resource works  [Conformance]
    /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:630
------------------------------
{"msg":"PASSED [sig-api-machinery] CustomResourceDefinition resources [Privileged:ClusterAdmin] Simple CustomResourceDefinition getting/updating/patching custom resource definition status sub-resource works  [Conformance]","total":-1,"completed":13,"skipped":36,"failed":0}

SSS
------------------------------
[BeforeEach] [Testpattern: Dynamic PV (immediate binding)] topology
  /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/storage/framework/testsuite.go:51
Jul 18 19:55:36.883: INFO: Only supported for providers [azure] (not skeleton)
... skipping 5 lines ...
[sig-storage] In-tree Volumes
/workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/storage/utils/framework.go:23
  [Driver: azure-disk]
  /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/storage/in_tree_volumes.go:58
    [Testpattern: Dynamic PV (immediate binding)] topology
    /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/storage/framework/testsuite.go:50
      should fail to schedule a pod which has topologies that conflict with AllowedTopologies [BeforeEach]
      /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/storage/testsuites/topology.go:192

      Only supported for providers [azure] (not skeleton)

      /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/storage/drivers/in_tree.go:1566
------------------------------
... skipping 68 lines ...
  /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/storage/in_tree_volumes.go:58
    [Testpattern: Pre-provisioned PV (filesystem volmode)] volumeMode
    /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/storage/framework/testsuite.go:50
      should not mount / map unused volumes in a pod [LinuxOnly]
      /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/storage/testsuites/volumemode.go:351
------------------------------
{"msg":"PASSED [sig-storage] In-tree Volumes [Driver: local][LocalVolumeType: block] [Testpattern: Pre-provisioned PV (filesystem volmode)] volumeMode should not mount / map unused volumes in a pod [LinuxOnly]","total":-1,"completed":8,"skipped":72,"failed":0}
[BeforeEach] [Testpattern: Pre-provisioned PV (default fs)] volumes
  /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/storage/framework/testsuite.go:51
Jul 18 19:55:36.908: INFO: Driver csi-hostpath doesn't support PreprovisionedPV -- skipping
[AfterEach] [Testpattern: Pre-provisioned PV (default fs)] volumes
  /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:186

... skipping 72 lines ...
STEP: Destroying namespace "services-5910" for this suite.
[AfterEach] [sig-network] Services
  /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/network/service.go:750

•
------------------------------
{"msg":"PASSED [sig-network] Services should provide secure master service  [Conformance]","total":-1,"completed":9,"skipped":77,"failed":0}
[BeforeEach] [Testpattern: Dynamic PV (default fs)] fsgroupchangepolicy
  /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/storage/framework/testsuite.go:51
Jul 18 19:55:37.202: INFO: Driver hostPath doesn't support DynamicPV -- skipping
[AfterEach] [Testpattern: Dynamic PV (default fs)] fsgroupchangepolicy
  /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:186

... skipping 108 lines ...
Jul 18 19:54:36.834: INFO: PersistentVolumeClaim pvc-tgdqm found but phase is Pending instead of Bound.
Jul 18 19:54:38.866: INFO: PersistentVolumeClaim pvc-tgdqm found and phase=Bound (2.063923176s)
STEP: Deleting the previously created pod
Jul 18 19:55:03.029: INFO: Deleting pod "pvc-volume-tester-x6h8v" in namespace "csi-mock-volumes-8569"
Jul 18 19:55:03.065: INFO: Wait up to 5m0s for pod "pvc-volume-tester-x6h8v" to be fully deleted
STEP: Checking CSI driver logs
Jul 18 19:55:09.165: INFO: Found NodeUnpublishVolume: {json: {"Method":"/csi.v1.Node/NodeUnpublishVolume","Request":{"volume_id":"4","target_path":"/var/lib/kubelet/pods/56dfadd8-6861-4f1a-abad-886bddf3a71a/volumes/kubernetes.io~csi/pvc-e7c53b57-f751-48d5-be9f-4735277bcb81/mount"},"Response":{},"Error":"","FullError":null} Method:NodeUnpublishVolume Request:{VolumeContext:map[]} FullError:{Code:OK Message:} Error:}
STEP: Deleting pod pvc-volume-tester-x6h8v
Jul 18 19:55:09.165: INFO: Deleting pod "pvc-volume-tester-x6h8v" in namespace "csi-mock-volumes-8569"
STEP: Deleting claim pvc-tgdqm
Jul 18 19:55:09.264: INFO: Waiting up to 2m0s for PersistentVolume pvc-e7c53b57-f751-48d5-be9f-4735277bcb81 to get deleted
Jul 18 19:55:09.294: INFO: PersistentVolume pvc-e7c53b57-f751-48d5-be9f-4735277bcb81 found and phase=Released (30.088457ms)
Jul 18 19:55:11.325: INFO: PersistentVolume pvc-e7c53b57-f751-48d5-be9f-4735277bcb81 found and phase=Released (2.061742323s)
... skipping 45 lines ...
/workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/storage/utils/framework.go:23
  CSIServiceAccountToken
  /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/storage/csi_mock_volume.go:1374
    token should not be plumbed down when CSIDriver is not deployed
    /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/storage/csi_mock_volume.go:1402
------------------------------
{"msg":"PASSED [sig-storage] CSI mock volume CSIServiceAccountToken token should not be plumbed down when CSIDriver is not deployed","total":-1,"completed":10,"skipped":50,"failed":0}

S
------------------------------
[BeforeEach] [Testpattern: Inline-volume (ext3)] volumes
  /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/storage/framework/testsuite.go:51
Jul 18 19:55:37.406: INFO: Driver hostPathSymlink doesn't support ext3 -- skipping
... skipping 25 lines ...
Jul 18 19:55:04.086: INFO: >>> kubeConfig: /root/.kube/config
STEP: Building a namespace api object, basename provisioning
STEP: Waiting for a default service account to be provisioned in namespace
[It] should support existing single file [LinuxOnly]
  /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/storage/testsuites/subpath.go:219
Jul 18 19:55:04.250: INFO: In-tree plugin kubernetes.io/host-path is not migrated, not validating any metrics
Jul 18 19:55:04.320: INFO: Waiting up to 5m0s for pod "hostpath-symlink-prep-provisioning-9246" in namespace "provisioning-9246" to be "Succeeded or Failed"
Jul 18 19:55:04.352: INFO: Pod "hostpath-symlink-prep-provisioning-9246": Phase="Pending", Reason="", readiness=false. Elapsed: 31.36374ms
Jul 18 19:55:06.384: INFO: Pod "hostpath-symlink-prep-provisioning-9246": Phase="Pending", Reason="", readiness=false. Elapsed: 2.063924377s
Jul 18 19:55:08.417: INFO: Pod "hostpath-symlink-prep-provisioning-9246": Phase="Pending", Reason="", readiness=false. Elapsed: 4.097019497s
Jul 18 19:55:10.450: INFO: Pod "hostpath-symlink-prep-provisioning-9246": Phase="Pending", Reason="", readiness=false. Elapsed: 6.129508356s
Jul 18 19:55:12.484: INFO: Pod "hostpath-symlink-prep-provisioning-9246": Phase="Pending", Reason="", readiness=false. Elapsed: 8.163216777s
Jul 18 19:55:14.515: INFO: Pod "hostpath-symlink-prep-provisioning-9246": Phase="Pending", Reason="", readiness=false. Elapsed: 10.194668707s
Jul 18 19:55:16.549: INFO: Pod "hostpath-symlink-prep-provisioning-9246": Phase="Pending", Reason="", readiness=false. Elapsed: 12.228329364s
Jul 18 19:55:18.585: INFO: Pod "hostpath-symlink-prep-provisioning-9246": Phase="Succeeded", Reason="", readiness=false. Elapsed: 14.265054111s
STEP: Saw pod success
Jul 18 19:55:18.586: INFO: Pod "hostpath-symlink-prep-provisioning-9246" satisfied condition "Succeeded or Failed"
Jul 18 19:55:18.586: INFO: Deleting pod "hostpath-symlink-prep-provisioning-9246" in namespace "provisioning-9246"
Jul 18 19:55:18.632: INFO: Wait up to 5m0s for pod "hostpath-symlink-prep-provisioning-9246" to be fully deleted
Jul 18 19:55:18.663: INFO: Creating resource for inline volume
STEP: Creating pod pod-subpath-test-inlinevolume-kjhb
STEP: Creating a pod to test subpath
Jul 18 19:55:18.700: INFO: Waiting up to 5m0s for pod "pod-subpath-test-inlinevolume-kjhb" in namespace "provisioning-9246" to be "Succeeded or Failed"
Jul 18 19:55:18.732: INFO: Pod "pod-subpath-test-inlinevolume-kjhb": Phase="Pending", Reason="", readiness=false. Elapsed: 31.61506ms
Jul 18 19:55:20.763: INFO: Pod "pod-subpath-test-inlinevolume-kjhb": Phase="Pending", Reason="", readiness=false. Elapsed: 2.063165342s
Jul 18 19:55:22.795: INFO: Pod "pod-subpath-test-inlinevolume-kjhb": Phase="Pending", Reason="", readiness=false. Elapsed: 4.094561209s
Jul 18 19:55:24.827: INFO: Pod "pod-subpath-test-inlinevolume-kjhb": Phase="Succeeded", Reason="", readiness=false. Elapsed: 6.127451999s
STEP: Saw pod success
Jul 18 19:55:24.828: INFO: Pod "pod-subpath-test-inlinevolume-kjhb" satisfied condition "Succeeded or Failed"
Jul 18 19:55:24.858: INFO: Trying to get logs from node 10.116.0.7 pod pod-subpath-test-inlinevolume-kjhb container test-container-subpath-inlinevolume-kjhb: <nil>
STEP: delete the pod
Jul 18 19:55:24.940: INFO: Waiting for pod pod-subpath-test-inlinevolume-kjhb to disappear
Jul 18 19:55:24.971: INFO: Pod pod-subpath-test-inlinevolume-kjhb no longer exists
STEP: Deleting pod pod-subpath-test-inlinevolume-kjhb
Jul 18 19:55:24.971: INFO: Deleting pod "pod-subpath-test-inlinevolume-kjhb" in namespace "provisioning-9246"
STEP: Deleting pod
Jul 18 19:55:25.001: INFO: Deleting pod "pod-subpath-test-inlinevolume-kjhb" in namespace "provisioning-9246"
Jul 18 19:55:25.067: INFO: Waiting up to 5m0s for pod "hostpath-symlink-prep-provisioning-9246" in namespace "provisioning-9246" to be "Succeeded or Failed"
Jul 18 19:55:25.098: INFO: Pod "hostpath-symlink-prep-provisioning-9246": Phase="Pending", Reason="", readiness=false. Elapsed: 30.716266ms
Jul 18 19:55:27.164: INFO: Pod "hostpath-symlink-prep-provisioning-9246": Phase="Pending", Reason="", readiness=false. Elapsed: 2.096539892s
Jul 18 19:55:29.195: INFO: Pod "hostpath-symlink-prep-provisioning-9246": Phase="Pending", Reason="", readiness=false. Elapsed: 4.127614905s
Jul 18 19:55:31.227: INFO: Pod "hostpath-symlink-prep-provisioning-9246": Phase="Pending", Reason="", readiness=false. Elapsed: 6.15962885s
Jul 18 19:55:33.259: INFO: Pod "hostpath-symlink-prep-provisioning-9246": Phase="Pending", Reason="", readiness=false. Elapsed: 8.191792938s
Jul 18 19:55:35.291: INFO: Pod "hostpath-symlink-prep-provisioning-9246": Phase="Pending", Reason="", readiness=false. Elapsed: 10.223766105s
Jul 18 19:55:37.323: INFO: Pod "hostpath-symlink-prep-provisioning-9246": Phase="Succeeded", Reason="", readiness=false. Elapsed: 12.256181919s
STEP: Saw pod success
Jul 18 19:55:37.323: INFO: Pod "hostpath-symlink-prep-provisioning-9246" satisfied condition "Succeeded or Failed"
Jul 18 19:55:37.323: INFO: Deleting pod "hostpath-symlink-prep-provisioning-9246" in namespace "provisioning-9246"
Jul 18 19:55:37.365: INFO: Wait up to 5m0s for pod "hostpath-symlink-prep-provisioning-9246" to be fully deleted
[AfterEach] [Testpattern: Inline-volume (default fs)] subPath
  /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:186
Jul 18 19:55:37.395: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready
STEP: Destroying namespace "provisioning-9246" for this suite.
... skipping 6 lines ...
  /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/storage/in_tree_volumes.go:58
    [Testpattern: Inline-volume (default fs)] subPath
    /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/storage/framework/testsuite.go:50
      should support existing single file [LinuxOnly]
      /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/storage/testsuites/subpath.go:219
------------------------------
{"msg":"PASSED [sig-storage] In-tree Volumes [Driver: hostPathSymlink] [Testpattern: Inline-volume (default fs)] subPath should support existing single file [LinuxOnly]","total":-1,"completed":12,"skipped":97,"failed":0}

SSS
------------------------------
[BeforeEach] [Testpattern: Dynamic PV (default fs)(allowExpansion)] volume-expand
  /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/storage/framework/testsuite.go:51
Jul 18 19:55:37.509: INFO: Driver emptydir doesn't support DynamicPV -- skipping
... skipping 60 lines ...
      Only supported for providers [openstack] (not skeleton)

      /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/storage/drivers/in_tree.go:1092
------------------------------
SSSSSSSSS
------------------------------
{"msg":"PASSED [sig-network] Conntrack should drop INVALID conntrack entries","total":-1,"completed":6,"skipped":32,"failed":0}
[BeforeEach] [Testpattern: Pre-provisioned PV (default fs)] subPath
  /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/storage/framework/testsuite.go:51
[BeforeEach] [Testpattern: Pre-provisioned PV (default fs)] subPath
  /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:185
STEP: Creating a kubernetes client
Jul 18 19:54:52.781: INFO: >>> kubeConfig: /root/.kube/config
... skipping 23 lines ...
Jul 18 19:55:20.512: INFO: PersistentVolumeClaim pvc-wxpd7 found but phase is Pending instead of Bound.
Jul 18 19:55:22.632: INFO: PersistentVolumeClaim pvc-wxpd7 found and phase=Bound (14.424203516s)
Jul 18 19:55:22.632: INFO: Waiting up to 3m0s for PersistentVolume local-vq4pm to have phase Bound
Jul 18 19:55:22.670: INFO: PersistentVolume local-vq4pm found and phase=Bound (37.822365ms)
STEP: Creating pod pod-subpath-test-preprovisionedpv-pr54
STEP: Creating a pod to test subpath
Jul 18 19:55:22.768: INFO: Waiting up to 5m0s for pod "pod-subpath-test-preprovisionedpv-pr54" in namespace "provisioning-4382" to be "Succeeded or Failed"
Jul 18 19:55:22.803: INFO: Pod "pod-subpath-test-preprovisionedpv-pr54": Phase="Pending", Reason="", readiness=false. Elapsed: 34.463006ms
Jul 18 19:55:24.836: INFO: Pod "pod-subpath-test-preprovisionedpv-pr54": Phase="Pending", Reason="", readiness=false. Elapsed: 2.067996764s
Jul 18 19:55:26.868: INFO: Pod "pod-subpath-test-preprovisionedpv-pr54": Phase="Pending", Reason="", readiness=false. Elapsed: 4.099569608s
Jul 18 19:55:28.903: INFO: Pod "pod-subpath-test-preprovisionedpv-pr54": Phase="Pending", Reason="", readiness=false. Elapsed: 6.134470808s
Jul 18 19:55:30.935: INFO: Pod "pod-subpath-test-preprovisionedpv-pr54": Phase="Pending", Reason="", readiness=false. Elapsed: 8.166799526s
Jul 18 19:55:32.971: INFO: Pod "pod-subpath-test-preprovisionedpv-pr54": Phase="Pending", Reason="", readiness=false. Elapsed: 10.202059351s
Jul 18 19:55:35.003: INFO: Pod "pod-subpath-test-preprovisionedpv-pr54": Phase="Pending", Reason="", readiness=false. Elapsed: 12.23449712s
Jul 18 19:55:37.035: INFO: Pod "pod-subpath-test-preprovisionedpv-pr54": Phase="Pending", Reason="", readiness=false. Elapsed: 14.266570527s
Jul 18 19:55:39.068: INFO: Pod "pod-subpath-test-preprovisionedpv-pr54": Phase="Succeeded", Reason="", readiness=false. Elapsed: 16.29973663s
STEP: Saw pod success
Jul 18 19:55:39.068: INFO: Pod "pod-subpath-test-preprovisionedpv-pr54" satisfied condition "Succeeded or Failed"
Jul 18 19:55:39.100: INFO: Trying to get logs from node 10.116.0.5 pod pod-subpath-test-preprovisionedpv-pr54 container test-container-volume-preprovisionedpv-pr54: <nil>
STEP: delete the pod
Jul 18 19:55:39.183: INFO: Waiting for pod pod-subpath-test-preprovisionedpv-pr54 to disappear
Jul 18 19:55:39.214: INFO: Pod pod-subpath-test-preprovisionedpv-pr54 no longer exists
STEP: Deleting pod pod-subpath-test-preprovisionedpv-pr54
Jul 18 19:55:39.214: INFO: Deleting pod "pod-subpath-test-preprovisionedpv-pr54" in namespace "provisioning-4382"
... skipping 36 lines ...
STEP: Building a namespace api object, basename security-context-test
STEP: Waiting for a default service account to be provisioned in namespace
[BeforeEach] [sig-node] Security Context
  /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/common/node/security_context.go:46
[It] should not allow privilege escalation when false [LinuxOnly] [NodeConformance] [Conformance]
  /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:630
Jul 18 19:55:30.993: INFO: Waiting up to 5m0s for pod "alpine-nnp-false-66362c7a-5176-4c21-922b-b238a61bbab8" in namespace "security-context-test-3290" to be "Succeeded or Failed"
Jul 18 19:55:31.024: INFO: Pod "alpine-nnp-false-66362c7a-5176-4c21-922b-b238a61bbab8": Phase="Pending", Reason="", readiness=false. Elapsed: 30.982669ms
Jul 18 19:55:33.055: INFO: Pod "alpine-nnp-false-66362c7a-5176-4c21-922b-b238a61bbab8": Phase="Pending", Reason="", readiness=false. Elapsed: 2.062260932s
Jul 18 19:55:35.089: INFO: Pod "alpine-nnp-false-66362c7a-5176-4c21-922b-b238a61bbab8": Phase="Pending", Reason="", readiness=false. Elapsed: 4.095972194s
Jul 18 19:55:37.120: INFO: Pod "alpine-nnp-false-66362c7a-5176-4c21-922b-b238a61bbab8": Phase="Pending", Reason="", readiness=false. Elapsed: 6.127303965s
Jul 18 19:55:39.155: INFO: Pod "alpine-nnp-false-66362c7a-5176-4c21-922b-b238a61bbab8": Phase="Pending", Reason="", readiness=false. Elapsed: 8.162488187s
Jul 18 19:55:41.187: INFO: Pod "alpine-nnp-false-66362c7a-5176-4c21-922b-b238a61bbab8": Phase="Succeeded", Reason="", readiness=false. Elapsed: 10.194278822s
Jul 18 19:55:41.187: INFO: Pod "alpine-nnp-false-66362c7a-5176-4c21-922b-b238a61bbab8" satisfied condition "Succeeded or Failed"
[AfterEach] [sig-node] Security Context
  /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:186
Jul 18 19:55:41.221: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready
STEP: Destroying namespace "security-context-test-3290" for this suite.


... skipping 2 lines ...
/workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/common/node/framework.go:23
  when creating containers with AllowPrivilegeEscalation
  /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/common/node/security_context.go:296
    should not allow privilege escalation when false [LinuxOnly] [NodeConformance] [Conformance]
    /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:630
------------------------------
{"msg":"PASSED [sig-node] Security Context when creating containers with AllowPrivilegeEscalation should not allow privilege escalation when false [LinuxOnly] [NodeConformance] [Conformance]","total":-1,"completed":12,"skipped":104,"failed":0}

S
------------------------------
[BeforeEach] [Testpattern: Inline-volume (default fs)] subPath
  /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/storage/framework/testsuite.go:51
Jul 18 19:55:41.333: INFO: Driver local doesn't support InlineVolume -- skipping
... skipping 283 lines ...
/workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/storage/utils/framework.go:23
  storage capacity
  /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/storage/csi_mock_volume.go:900
    unlimited
    /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/storage/csi_mock_volume.go:958
------------------------------
{"msg":"PASSED [sig-storage] CSI mock volume storage capacity unlimited","total":-1,"completed":9,"skipped":65,"failed":0}

SSSS
------------------------------
[BeforeEach] [Testpattern: Dynamic PV (delayed binding)] topology
  /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/storage/framework/testsuite.go:51
Jul 18 19:55:45.811: INFO: Driver local doesn't support DynamicPV -- skipping
... skipping 5 lines ...
[sig-storage] In-tree Volumes
/workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/storage/utils/framework.go:23
  [Driver: local][LocalVolumeType: block]
  /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/storage/in_tree_volumes.go:58
    [Testpattern: Dynamic PV (delayed binding)] topology
    /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/storage/framework/testsuite.go:50
      should fail to schedule a pod which has topologies that conflict with AllowedTopologies [BeforeEach]
      /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/storage/testsuites/topology.go:192

      Driver local doesn't support DynamicPV -- skipping

      /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/storage/framework/testsuite.go:116
------------------------------
... skipping 19 lines ...
      Driver local doesn't support DynamicPV -- skipping

      /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/storage/framework/testsuite.go:116
------------------------------
SSSSSSSS
------------------------------
{"msg":"PASSED [sig-network] Networking Granular Checks: Pods should function for node-pod communication: http [LinuxOnly] [NodeConformance] [Conformance]","total":-1,"completed":8,"skipped":39,"failed":0}
[BeforeEach] [sig-storage] Downward API volume
  /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:185
STEP: Creating a kubernetes client
Jul 18 19:55:33.054: INFO: >>> kubeConfig: /root/.kube/config
STEP: Building a namespace api object, basename downward-api
STEP: Waiting for a default service account to be provisioned in namespace
... skipping 18 lines ...
• [SLOW TEST:13.610 seconds]
[sig-storage] Downward API volume
/workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/common/storage/framework.go:23
  should update labels on modification [NodeConformance] [Conformance]
  /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:630
------------------------------
{"msg":"PASSED [sig-storage] Downward API volume should update labels on modification [NodeConformance] [Conformance]","total":-1,"completed":9,"skipped":39,"failed":0}

SSSSS
------------------------------
[BeforeEach] [Testpattern: Pre-provisioned PV (default fs)] volumes
  /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/storage/framework/testsuite.go:51
Jul 18 19:55:46.694: INFO: Driver hostPathSymlink doesn't support PreprovisionedPV -- skipping
... skipping 78 lines ...
  /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/storage/persistent_volumes-local.go:192
    One pod requesting one prebound PVC
    /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/storage/persistent_volumes-local.go:209
      should be able to mount volume and write from pod1
      /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/storage/persistent_volumes-local.go:238
------------------------------
{"msg":"PASSED [sig-storage] PersistentVolumes-local  [Volume type: dir-link] One pod requesting one prebound PVC should be able to mount volume and write from pod1","total":-1,"completed":11,"skipped":93,"failed":0}

S
------------------------------
[BeforeEach] [sig-node] Variable Expansion
  /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:185
STEP: Creating a kubernetes client
Jul 18 19:55:41.383: INFO: >>> kubeConfig: /root/.kube/config
STEP: Building a namespace api object, basename var-expansion
STEP: Waiting for a default service account to be provisioned in namespace
[It] should allow substituting values in a container's args [NodeConformance] [Conformance]
  /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:630
STEP: Creating a pod to test substitution in container's args
Jul 18 19:55:41.580: INFO: Waiting up to 5m0s for pod "var-expansion-cf1caa55-a457-47fb-b69e-1180631a01bc" in namespace "var-expansion-7414" to be "Succeeded or Failed"
Jul 18 19:55:41.611: INFO: Pod "var-expansion-cf1caa55-a457-47fb-b69e-1180631a01bc": Phase="Pending", Reason="", readiness=false. Elapsed: 31.109324ms
Jul 18 19:55:43.644: INFO: Pod "var-expansion-cf1caa55-a457-47fb-b69e-1180631a01bc": Phase="Pending", Reason="", readiness=false. Elapsed: 2.064363346s
Jul 18 19:55:45.676: INFO: Pod "var-expansion-cf1caa55-a457-47fb-b69e-1180631a01bc": Phase="Pending", Reason="", readiness=false. Elapsed: 4.096164878s
Jul 18 19:55:47.709: INFO: Pod "var-expansion-cf1caa55-a457-47fb-b69e-1180631a01bc": Phase="Succeeded", Reason="", readiness=false. Elapsed: 6.129118976s
STEP: Saw pod success
Jul 18 19:55:47.709: INFO: Pod "var-expansion-cf1caa55-a457-47fb-b69e-1180631a01bc" satisfied condition "Succeeded or Failed"
Jul 18 19:55:47.741: INFO: Trying to get logs from node 10.116.0.5 pod var-expansion-cf1caa55-a457-47fb-b69e-1180631a01bc container dapi-container: <nil>
STEP: delete the pod
Jul 18 19:55:47.815: INFO: Waiting for pod var-expansion-cf1caa55-a457-47fb-b69e-1180631a01bc to disappear
Jul 18 19:55:47.846: INFO: Pod var-expansion-cf1caa55-a457-47fb-b69e-1180631a01bc no longer exists
[AfterEach] [sig-node] Variable Expansion
  /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:186
... skipping 4 lines ...
• [SLOW TEST:6.558 seconds]
[sig-node] Variable Expansion
/workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/common/node/framework.go:23
  should allow substituting values in a container's args [NodeConformance] [Conformance]
  /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:630
------------------------------
{"msg":"PASSED [sig-node] Variable Expansion should allow substituting values in a container's args [NodeConformance] [Conformance]","total":-1,"completed":13,"skipped":116,"failed":0}
[BeforeEach] [Testpattern: Inline-volume (ext4)] volumes
  /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/storage/framework/testsuite.go:51
Jul 18 19:55:47.952: INFO: Driver local doesn't support InlineVolume -- skipping
[AfterEach] [Testpattern: Inline-volume (ext4)] volumes
  /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:186

... skipping 129 lines ...
  /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:186
Jul 18 19:55:48.859: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready
STEP: Destroying namespace "proxy-6905" for this suite.

•
------------------------------
{"msg":"PASSED [sig-network] Proxy version v1 should proxy logs on node with explicit kubelet port using proxy subresource ","total":-1,"completed":12,"skipped":94,"failed":0}

SS
------------------------------
[BeforeEach] [sig-storage] PersistentVolumes-local 
  /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:185
STEP: Creating a kubernetes client
... skipping 138 lines ...
• [SLOW TEST:14.623 seconds]
[sig-api-machinery] CustomResourcePublishOpenAPI [Privileged:ClusterAdmin]
/workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/apimachinery/framework.go:23
  works for CRD preserving unknown fields in an embedded object [Conformance]
  /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:630
------------------------------
{"msg":"PASSED [sig-api-machinery] CustomResourcePublishOpenAPI [Privileged:ClusterAdmin] works for CRD preserving unknown fields in an embedded object [Conformance]","total":-1,"completed":13,"skipped":123,"failed":0}

SSSS
------------------------------
[BeforeEach] [Testpattern: Dynamic PV (ext4)] volumes
  /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/storage/framework/testsuite.go:51
Jul 18 19:55:52.253: INFO: Only supported for providers [azure] (not skeleton)
... skipping 69 lines ...
/workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/common/network/framework.go:23
  Granular Checks: Pods
  /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/common/network/networking.go:30
    should function for intra-pod communication: http [NodeConformance] [Conformance]
    /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:630
------------------------------
{"msg":"PASSED [sig-network] Networking Granular Checks: Pods should function for intra-pod communication: http [NodeConformance] [Conformance]","total":-1,"completed":5,"skipped":27,"failed":0}

S
------------------------------
[BeforeEach] [sig-storage] ConfigMap
  /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:185
STEP: Creating a kubernetes client
Jul 18 19:55:46.710: INFO: >>> kubeConfig: /root/.kube/config
STEP: Building a namespace api object, basename configmap
STEP: Waiting for a default service account to be provisioned in namespace
[It] should be consumable from pods in volume as non-root with defaultMode and fsGroup set [LinuxOnly] [NodeFeature:FSGroup]
  /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/common/storage/configmap_volume.go:59
STEP: Creating configMap with name configmap-test-volume-79755434-ca19-458a-8c9f-fd20df51f49e
STEP: Creating a pod to test consume configMaps
Jul 18 19:55:46.935: INFO: Waiting up to 5m0s for pod "pod-configmaps-0c86d227-ea73-4947-8a53-8551c5859cc2" in namespace "configmap-743" to be "Succeeded or Failed"
Jul 18 19:55:46.968: INFO: Pod "pod-configmaps-0c86d227-ea73-4947-8a53-8551c5859cc2": Phase="Pending", Reason="", readiness=false. Elapsed: 33.044306ms
Jul 18 19:55:49.000: INFO: Pod "pod-configmaps-0c86d227-ea73-4947-8a53-8551c5859cc2": Phase="Pending", Reason="", readiness=false. Elapsed: 2.064821405s
Jul 18 19:55:51.032: INFO: Pod "pod-configmaps-0c86d227-ea73-4947-8a53-8551c5859cc2": Phase="Pending", Reason="", readiness=false. Elapsed: 4.097051785s
Jul 18 19:55:53.065: INFO: Pod "pod-configmaps-0c86d227-ea73-4947-8a53-8551c5859cc2": Phase="Succeeded", Reason="", readiness=false. Elapsed: 6.129530276s
STEP: Saw pod success
Jul 18 19:55:53.065: INFO: Pod "pod-configmaps-0c86d227-ea73-4947-8a53-8551c5859cc2" satisfied condition "Succeeded or Failed"
Jul 18 19:55:53.095: INFO: Trying to get logs from node 10.116.0.7 pod pod-configmaps-0c86d227-ea73-4947-8a53-8551c5859cc2 container agnhost-container: <nil>
STEP: delete the pod
Jul 18 19:55:53.169: INFO: Waiting for pod pod-configmaps-0c86d227-ea73-4947-8a53-8551c5859cc2 to disappear
Jul 18 19:55:53.199: INFO: Pod pod-configmaps-0c86d227-ea73-4947-8a53-8551c5859cc2 no longer exists
[AfterEach] [sig-storage] ConfigMap
  /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:186
... skipping 4 lines ...
• [SLOW TEST:6.582 seconds]
[sig-storage] ConfigMap
/workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/common/storage/framework.go:23
  should be consumable from pods in volume as non-root with defaultMode and fsGroup set [LinuxOnly] [NodeFeature:FSGroup]
  /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/common/storage/configmap_volume.go:59
------------------------------
{"msg":"PASSED [sig-storage] ConfigMap should be consumable from pods in volume as non-root with defaultMode and fsGroup set [LinuxOnly] [NodeFeature:FSGroup]","total":-1,"completed":10,"skipped":45,"failed":0}

SSS
------------------------------
[BeforeEach] [sig-storage] Downward API volume
  /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:185
STEP: Creating a kubernetes client
... skipping 2 lines ...
STEP: Waiting for a default service account to be provisioned in namespace
[BeforeEach] [sig-storage] Downward API volume
  /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/common/storage/downwardapi_volume.go:41
[It] should provide container's cpu limit [NodeConformance] [Conformance]
  /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:630
STEP: Creating a pod to test downward API volume plugin
Jul 18 19:55:48.214: INFO: Waiting up to 5m0s for pod "downwardapi-volume-7d87dc41-147f-4573-bb99-3bea00000d66" in namespace "downward-api-4014" to be "Succeeded or Failed"
Jul 18 19:55:48.246: INFO: Pod "downwardapi-volume-7d87dc41-147f-4573-bb99-3bea00000d66": Phase="Pending", Reason="", readiness=false. Elapsed: 32.651022ms
Jul 18 19:55:50.280: INFO: Pod "downwardapi-volume-7d87dc41-147f-4573-bb99-3bea00000d66": Phase="Pending", Reason="", readiness=false. Elapsed: 2.066262612s
Jul 18 19:55:52.313: INFO: Pod "downwardapi-volume-7d87dc41-147f-4573-bb99-3bea00000d66": Phase="Pending", Reason="", readiness=false. Elapsed: 4.09919165s
Jul 18 19:55:54.345: INFO: Pod "downwardapi-volume-7d87dc41-147f-4573-bb99-3bea00000d66": Phase="Pending", Reason="", readiness=false. Elapsed: 6.131075869s
Jul 18 19:55:56.378: INFO: Pod "downwardapi-volume-7d87dc41-147f-4573-bb99-3bea00000d66": Phase="Succeeded", Reason="", readiness=false. Elapsed: 8.164209461s
STEP: Saw pod success
Jul 18 19:55:56.378: INFO: Pod "downwardapi-volume-7d87dc41-147f-4573-bb99-3bea00000d66" satisfied condition "Succeeded or Failed"
Jul 18 19:55:56.409: INFO: Trying to get logs from node 10.116.0.5 pod downwardapi-volume-7d87dc41-147f-4573-bb99-3bea00000d66 container client-container: <nil>
STEP: delete the pod
Jul 18 19:55:56.492: INFO: Waiting for pod downwardapi-volume-7d87dc41-147f-4573-bb99-3bea00000d66 to disappear
Jul 18 19:55:56.523: INFO: Pod downwardapi-volume-7d87dc41-147f-4573-bb99-3bea00000d66 no longer exists
[AfterEach] [sig-storage] Downward API volume
  /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:186
... skipping 4 lines ...
• [SLOW TEST:8.603 seconds]
[sig-storage] Downward API volume
/workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/common/storage/framework.go:23
  should provide container's cpu limit [NodeConformance] [Conformance]
  /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:630
------------------------------
{"msg":"PASSED [sig-storage] Downward API volume should provide container's cpu limit [NodeConformance] [Conformance]","total":-1,"completed":14,"skipped":131,"failed":0}

SS
------------------------------
[BeforeEach] [Testpattern: Inline-volume (default fs)] subPath
  /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/storage/framework/testsuite.go:51
Jul 18 19:55:56.640: INFO: Driver local doesn't support InlineVolume -- skipping
... skipping 79 lines ...
      Driver local doesn't support DynamicPV -- skipping

      /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/storage/framework/testsuite.go:116
------------------------------
S
------------------------------
{"msg":"PASSED [sig-storage] In-tree Volumes [Driver: local][LocalVolumeType: blockfs] [Testpattern: Pre-provisioned PV (default fs)] subPath should support non-existent path","total":-1,"completed":7,"skipped":32,"failed":0}
[BeforeEach] [sig-api-machinery] ResourceQuota
  /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:185
STEP: Creating a kubernetes client
Jul 18 19:55:40.533: INFO: >>> kubeConfig: /root/.kube/config
STEP: Building a namespace api object, basename resourcequota
STEP: Waiting for a default service account to be provisioned in namespace
... skipping 22 lines ...
• [SLOW TEST:16.743 seconds]
[sig-api-machinery] ResourceQuota
/workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/apimachinery/framework.go:23
  should verify ResourceQuota with best effort scope. [Conformance]
  /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:630
------------------------------
{"msg":"PASSED [sig-api-machinery] ResourceQuota should verify ResourceQuota with best effort scope. [Conformance]","total":-1,"completed":8,"skipped":32,"failed":0}

SSS
------------------------------
[BeforeEach] [sig-apps] CronJob
  /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:185
STEP: Creating a kubernetes client
... skipping 20 lines ...
• [SLOW TEST:64.480 seconds]
[sig-apps] CronJob
/workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/apps/framework.go:23
  should replace jobs when ReplaceConcurrent [Conformance]
  /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:630
------------------------------
{"msg":"PASSED [sig-apps] CronJob should replace jobs when ReplaceConcurrent [Conformance]","total":-1,"completed":8,"skipped":28,"failed":0}

SS
------------------------------
[BeforeEach] [sig-storage] PersistentVolumes-local 
  /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:185
STEP: Creating a kubernetes client
... skipping 77 lines ...
STEP: Waiting for a default service account to be provisioned in namespace
[BeforeEach] [sig-storage] Projected downwardAPI
  /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/common/storage/projected_downwardapi.go:41
[It] should provide container's memory limit [NodeConformance] [Conformance]
  /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:630
STEP: Creating a pod to test downward API volume plugin
Jul 18 19:55:52.871: INFO: Waiting up to 5m0s for pod "downwardapi-volume-e7366bef-5715-4b4f-9b2d-779efe6e4582" in namespace "projected-7765" to be "Succeeded or Failed"
Jul 18 19:55:52.903: INFO: Pod "downwardapi-volume-e7366bef-5715-4b4f-9b2d-779efe6e4582": Phase="Pending", Reason="", readiness=false. Elapsed: 32.334427ms
Jul 18 19:55:54.937: INFO: Pod "downwardapi-volume-e7366bef-5715-4b4f-9b2d-779efe6e4582": Phase="Pending", Reason="", readiness=false. Elapsed: 2.065722094s
Jul 18 19:55:56.969: INFO: Pod "downwardapi-volume-e7366bef-5715-4b4f-9b2d-779efe6e4582": Phase="Pending", Reason="", readiness=false. Elapsed: 4.098553083s
Jul 18 19:55:59.002: INFO: Pod "downwardapi-volume-e7366bef-5715-4b4f-9b2d-779efe6e4582": Phase="Pending", Reason="", readiness=false. Elapsed: 6.131120055s
Jul 18 19:56:01.043: INFO: Pod "downwardapi-volume-e7366bef-5715-4b4f-9b2d-779efe6e4582": Phase="Pending", Reason="", readiness=false. Elapsed: 8.17248501s
Jul 18 19:56:03.076: INFO: Pod "downwardapi-volume-e7366bef-5715-4b4f-9b2d-779efe6e4582": Phase="Succeeded", Reason="", readiness=false. Elapsed: 10.20540221s
STEP: Saw pod success
Jul 18 19:56:03.076: INFO: Pod "downwardapi-volume-e7366bef-5715-4b4f-9b2d-779efe6e4582" satisfied condition "Succeeded or Failed"
Jul 18 19:56:03.108: INFO: Trying to get logs from node 10.116.0.5 pod downwardapi-volume-e7366bef-5715-4b4f-9b2d-779efe6e4582 container client-container: <nil>
STEP: delete the pod
Jul 18 19:56:03.184: INFO: Waiting for pod downwardapi-volume-e7366bef-5715-4b4f-9b2d-779efe6e4582 to disappear
Jul 18 19:56:03.216: INFO: Pod downwardapi-volume-e7366bef-5715-4b4f-9b2d-779efe6e4582 no longer exists
[AfterEach] [sig-storage] Projected downwardAPI
  /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:186
... skipping 4 lines ...
• [SLOW TEST:10.637 seconds]
[sig-storage] Projected downwardAPI
/workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/common/storage/framework.go:23
  should provide container's memory limit [NodeConformance] [Conformance]
  /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:630
------------------------------
{"msg":"PASSED [sig-storage] Projected downwardAPI should provide container's memory limit [NodeConformance] [Conformance]","total":-1,"completed":6,"skipped":28,"failed":0}

SS
------------------------------
[BeforeEach] [sig-cli] Kubectl client
  /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:185
STEP: Creating a kubernetes client
... skipping 232 lines ...
Jul 18 19:55:51.545: INFO: PersistentVolumeClaim pvc-p2k6j found but phase is Pending instead of Bound.
Jul 18 19:55:53.576: INFO: PersistentVolumeClaim pvc-p2k6j found and phase=Bound (16.284024191s)
Jul 18 19:55:53.576: INFO: Waiting up to 3m0s for PersistentVolume local-tf4rt to have phase Bound
Jul 18 19:55:53.607: INFO: PersistentVolume local-tf4rt found and phase=Bound (30.726762ms)
STEP: Creating pod pod-subpath-test-preprovisionedpv-8zvn
STEP: Creating a pod to test subpath
Jul 18 19:55:53.707: INFO: Waiting up to 5m0s for pod "pod-subpath-test-preprovisionedpv-8zvn" in namespace "provisioning-9950" to be "Succeeded or Failed"
Jul 18 19:55:53.738: INFO: Pod "pod-subpath-test-preprovisionedpv-8zvn": Phase="Pending", Reason="", readiness=false. Elapsed: 30.774284ms
Jul 18 19:55:55.774: INFO: Pod "pod-subpath-test-preprovisionedpv-8zvn": Phase="Pending", Reason="", readiness=false. Elapsed: 2.06677714s
Jul 18 19:55:57.805: INFO: Pod "pod-subpath-test-preprovisionedpv-8zvn": Phase="Pending", Reason="", readiness=false. Elapsed: 4.097896145s
Jul 18 19:55:59.836: INFO: Pod "pod-subpath-test-preprovisionedpv-8zvn": Phase="Pending", Reason="", readiness=false. Elapsed: 6.128573784s
Jul 18 19:56:01.867: INFO: Pod "pod-subpath-test-preprovisionedpv-8zvn": Phase="Pending", Reason="", readiness=false. Elapsed: 8.159998716s
Jul 18 19:56:03.899: INFO: Pod "pod-subpath-test-preprovisionedpv-8zvn": Phase="Pending", Reason="", readiness=false. Elapsed: 10.191715716s
Jul 18 19:56:05.932: INFO: Pod "pod-subpath-test-preprovisionedpv-8zvn": Phase="Pending", Reason="", readiness=false. Elapsed: 12.224240253s
Jul 18 19:56:07.964: INFO: Pod "pod-subpath-test-preprovisionedpv-8zvn": Phase="Succeeded", Reason="", readiness=false. Elapsed: 14.25683988s
STEP: Saw pod success
Jul 18 19:56:07.964: INFO: Pod "pod-subpath-test-preprovisionedpv-8zvn" satisfied condition "Succeeded or Failed"
Jul 18 19:56:07.995: INFO: Trying to get logs from node 10.116.0.5 pod pod-subpath-test-preprovisionedpv-8zvn container test-container-subpath-preprovisionedpv-8zvn: <nil>
STEP: delete the pod
Jul 18 19:56:08.071: INFO: Waiting for pod pod-subpath-test-preprovisionedpv-8zvn to disappear
Jul 18 19:56:08.103: INFO: Pod pod-subpath-test-preprovisionedpv-8zvn no longer exists
STEP: Deleting pod pod-subpath-test-preprovisionedpv-8zvn
Jul 18 19:56:08.103: INFO: Deleting pod "pod-subpath-test-preprovisionedpv-8zvn" in namespace "provisioning-9950"
... skipping 24 lines ...
  /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/storage/in_tree_volumes.go:58
    [Testpattern: Pre-provisioned PV (default fs)] subPath
    /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/storage/framework/testsuite.go:50
      should support readOnly file specified in the volumeMount [LinuxOnly]
      /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/storage/testsuites/subpath.go:379
------------------------------
{"msg":"PASSED [sig-storage] In-tree Volumes [Driver: local][LocalVolumeType: block] [Testpattern: Pre-provisioned PV (default fs)] subPath should support readOnly file specified in the volumeMount [LinuxOnly]","total":-1,"completed":6,"skipped":69,"failed":0}

SSSSS
------------------------------
[BeforeEach] [Testpattern: Dynamic PV (default fs)] volumes
  /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/storage/framework/testsuite.go:51
Jul 18 19:56:09.158: INFO: Driver emptydir doesn't support DynamicPV -- skipping
... skipping 93 lines ...
/workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/common/network/framework.go:23
  Granular Checks: Pods
  /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/common/network/networking.go:30
    should function for intra-pod communication: udp [NodeConformance] [Conformance]
    /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:630
------------------------------
{"msg":"PASSED [sig-network] Networking Granular Checks: Pods should function for intra-pod communication: udp [NodeConformance] [Conformance]","total":-1,"completed":14,"skipped":101,"failed":0}

SSSSS
------------------------------
[BeforeEach] [Testpattern: Dynamic PV (block volmode)(allowExpansion)] volume-expand
  /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/storage/framework/testsuite.go:51
Jul 18 19:56:12.112: INFO: Driver local doesn't support DynamicPV -- skipping
... skipping 82 lines ...
Jul 18 19:55:49.932: INFO: PersistentVolumeClaim pvc-8gtkq found but phase is Pending instead of Bound.
Jul 18 19:55:51.963: INFO: PersistentVolumeClaim pvc-8gtkq found and phase=Bound (6.128102487s)
Jul 18 19:55:51.963: INFO: Waiting up to 3m0s for PersistentVolume local-57mvg to have phase Bound
Jul 18 19:55:51.994: INFO: PersistentVolume local-57mvg found and phase=Bound (30.685244ms)
STEP: Creating pod pod-subpath-test-preprovisionedpv-xm5c
STEP: Creating a pod to test subpath
Jul 18 19:55:52.093: INFO: Waiting up to 5m0s for pod "pod-subpath-test-preprovisionedpv-xm5c" in namespace "provisioning-1322" to be "Succeeded or Failed"
Jul 18 19:55:52.124: INFO: Pod "pod-subpath-test-preprovisionedpv-xm5c": Phase="Pending", Reason="", readiness=false. Elapsed: 30.743156ms
Jul 18 19:55:54.157: INFO: Pod "pod-subpath-test-preprovisionedpv-xm5c": Phase="Pending", Reason="", readiness=false. Elapsed: 2.064352758s
Jul 18 19:55:56.190: INFO: Pod "pod-subpath-test-preprovisionedpv-xm5c": Phase="Pending", Reason="", readiness=false. Elapsed: 4.09717072s
Jul 18 19:55:58.223: INFO: Pod "pod-subpath-test-preprovisionedpv-xm5c": Phase="Pending", Reason="", readiness=false. Elapsed: 6.12981444s
Jul 18 19:56:00.255: INFO: Pod "pod-subpath-test-preprovisionedpv-xm5c": Phase="Pending", Reason="", readiness=false. Elapsed: 8.161906965s
Jul 18 19:56:02.287: INFO: Pod "pod-subpath-test-preprovisionedpv-xm5c": Phase="Pending", Reason="", readiness=false. Elapsed: 10.194567721s
Jul 18 19:56:04.320: INFO: Pod "pod-subpath-test-preprovisionedpv-xm5c": Phase="Pending", Reason="", readiness=false. Elapsed: 12.227221564s
Jul 18 19:56:06.353: INFO: Pod "pod-subpath-test-preprovisionedpv-xm5c": Phase="Pending", Reason="", readiness=false. Elapsed: 14.259952646s
Jul 18 19:56:08.386: INFO: Pod "pod-subpath-test-preprovisionedpv-xm5c": Phase="Pending", Reason="", readiness=false. Elapsed: 16.293407159s
Jul 18 19:56:10.419: INFO: Pod "pod-subpath-test-preprovisionedpv-xm5c": Phase="Pending", Reason="", readiness=false. Elapsed: 18.326412952s
Jul 18 19:56:12.453: INFO: Pod "pod-subpath-test-preprovisionedpv-xm5c": Phase="Succeeded", Reason="", readiness=false. Elapsed: 20.360032093s
STEP: Saw pod success
Jul 18 19:56:12.453: INFO: Pod "pod-subpath-test-preprovisionedpv-xm5c" satisfied condition "Succeeded or Failed"
Jul 18 19:56:12.485: INFO: Trying to get logs from node 10.116.0.7 pod pod-subpath-test-preprovisionedpv-xm5c container test-container-subpath-preprovisionedpv-xm5c: <nil>
STEP: delete the pod
Jul 18 19:56:12.561: INFO: Waiting for pod pod-subpath-test-preprovisionedpv-xm5c to disappear
Jul 18 19:56:12.593: INFO: Pod pod-subpath-test-preprovisionedpv-xm5c no longer exists
STEP: Deleting pod pod-subpath-test-preprovisionedpv-xm5c
Jul 18 19:56:12.593: INFO: Deleting pod "pod-subpath-test-preprovisionedpv-xm5c" in namespace "provisioning-1322"
... skipping 19 lines ...
  /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/storage/in_tree_volumes.go:58
    [Testpattern: Pre-provisioned PV (default fs)] subPath
    /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/storage/framework/testsuite.go:50
      should support readOnly directory specified in the volumeMount
      /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/storage/testsuites/subpath.go:364
------------------------------
{"msg":"PASSED [sig-storage] In-tree Volumes [Driver: local][LocalVolumeType: dir-bindmounted] [Testpattern: Pre-provisioned PV (default fs)] subPath should support readOnly directory specified in the volumeMount","total":-1,"completed":10,"skipped":84,"failed":0}

SSSSS
------------------------------
[BeforeEach] [Testpattern: Pre-provisioned PV (ext3)] volumes
  /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/storage/framework/testsuite.go:51
Jul 18 19:56:13.142: INFO: Only supported for providers [openstack] (not skeleton)
... skipping 72 lines ...
• [SLOW TEST:92.541 seconds]
[sig-storage] Projected configMap
/workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/common/storage/framework.go:23
  updates should be reflected in volume [NodeConformance] [Conformance]
  /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:630
------------------------------
{"msg":"PASSED [sig-storage] Projected configMap updates should be reflected in volume [NodeConformance] [Conformance]","total":-1,"completed":10,"skipped":81,"failed":0}
[BeforeEach] [Testpattern: Dynamic PV (default fs)] volumes
  /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/storage/framework/testsuite.go:51
Jul 18 19:56:13.681: INFO: Driver local doesn't support DynamicPV -- skipping
[AfterEach] [Testpattern: Dynamic PV (default fs)] volumes
  /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:186

... skipping 121 lines ...
/workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/storage/utils/framework.go:23
  CSI attach test using mock driver
  /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/storage/csi_mock_volume.go:316
    should preserve attachment policy when no CSIDriver present
    /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/storage/csi_mock_volume.go:338
------------------------------
{"msg":"PASSED [sig-storage] CSI mock volume CSI attach test using mock driver should preserve attachment policy when no CSIDriver present","total":-1,"completed":10,"skipped":41,"failed":0}

SSSS
------------------------------
{"msg":"PASSED [sig-storage] PersistentVolumes-local  [Volume type: blockfswithformat] One pod requesting one prebound PVC should be able to mount volume and read from pod1","total":-1,"completed":11,"skipped":79,"failed":0}
[BeforeEach] [sig-node] Container Runtime
  /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:185
STEP: Creating a kubernetes client
Jul 18 19:56:02.668: INFO: >>> kubeConfig: /root/.kube/config
STEP: Building a namespace api object, basename container-runtime
STEP: Waiting for a default service account to be provisioned in namespace
... skipping 19 lines ...
  /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/common/node/runtime.go:41
    on terminated container
    /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/common/node/runtime.go:134
      should report termination message [LinuxOnly] if TerminationMessagePath is set as non-root user and at a non-default path [NodeConformance] [Conformance]
      /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:630
------------------------------
{"msg":"PASSED [sig-node] Container Runtime blackbox test on terminated container should report termination message [LinuxOnly] if TerminationMessagePath is set as non-root user and at a non-default path [NodeConformance] [Conformance]","total":-1,"completed":12,"skipped":79,"failed":0}

SSSSSS
------------------------------
[BeforeEach] [Testpattern: Dynamic PV (default fs)] subPath
  /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/storage/framework/testsuite.go:51
Jul 18 19:56:14.478: INFO: Driver hostPathSymlink doesn't support DynamicPV -- skipping
... skipping 60 lines ...
• [SLOW TEST:253.137 seconds]
[sig-node] Probing container
/workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/common/node/framework.go:23
  should *not* be restarted with a tcp:8080 liveness probe [NodeConformance] [Conformance]
  /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:630
------------------------------
{"msg":"PASSED [sig-node] Probing container should *not* be restarted with a tcp:8080 liveness probe [NodeConformance] [Conformance]","total":-1,"completed":4,"skipped":18,"failed":0}

SSS
------------------------------
[BeforeEach] [sig-node] Container Runtime
  /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:185
STEP: Creating a kubernetes client
... skipping 22 lines ...
  /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/common/node/runtime.go:41
    on terminated container
    /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/common/node/runtime.go:134
      should report termination message [LinuxOnly] from file when pod succeeds and TerminationMessagePolicy FallbackToLogsOnError is set [NodeConformance] [Conformance]
      /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:630
------------------------------
{"msg":"PASSED [sig-node] Container Runtime blackbox test on terminated container should report termination message [LinuxOnly] from file when pod succeeds and TerminationMessagePolicy FallbackToLogsOnError is set [NodeConformance] [Conformance]","total":-1,"completed":15,"skipped":141,"failed":0}

SSS
------------------------------
[BeforeEach] [sig-apps] DisruptionController
  /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:185
STEP: Creating a kubernetes client
... skipping 21 lines ...
• [SLOW TEST:18.599 seconds]
[sig-apps] DisruptionController
/workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/apps/framework.go:23
  evictions: maxUnavailable allow single eviction, percentage => should allow an eviction
  /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/apps/disruption.go:267
------------------------------
{"msg":"PASSED [sig-apps] DisruptionController evictions: maxUnavailable allow single eviction, percentage =\u003e should allow an eviction","total":-1,"completed":9,"skipped":35,"failed":0}

SS
------------------------------
[BeforeEach] [Testpattern: Dynamic PV (default fs)(allowExpansion)] volume-expand
  /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/storage/framework/testsuite.go:51
Jul 18 19:56:15.912: INFO: Driver local doesn't support DynamicPV -- skipping
... skipping 28 lines ...
[sig-storage] CSI Volumes
/workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/storage/utils/framework.go:23
  [Driver: csi-hostpath]
  /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/storage/csi_volumes.go:40
    [Testpattern: Dynamic PV (immediate binding)] topology
    /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/storage/framework/testsuite.go:50
      should fail to schedule a pod which has topologies that conflict with AllowedTopologies [BeforeEach]
      /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/storage/testsuites/topology.go:192

      Driver "csi-hostpath" does not support topology - skipping

      /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/storage/testsuites/topology.go:92
------------------------------
S
------------------------------
{"msg":"PASSED [sig-storage] Projected configMap should be consumable in multiple volumes in the same pod [NodeConformance] [Conformance]","total":-1,"completed":9,"skipped":115,"failed":0}
[BeforeEach] [Testpattern: Dynamic PV (block volmode)] volumes
  /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/storage/framework/testsuite.go:51
[BeforeEach] [Testpattern: Dynamic PV (block volmode)] volumes
  /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:185
STEP: Creating a kubernetes client
Jul 18 19:54:05.636: INFO: >>> kubeConfig: /root/.kube/config
... skipping 152 lines ...
  /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/storage/csi_volumes.go:40
    [Testpattern: Dynamic PV (block volmode)] volumes
    /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/storage/framework/testsuite.go:50
      should store data
      /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/storage/testsuites/volumes.go:159
------------------------------
{"msg":"PASSED [sig-node] Secrets should patch a secret [Conformance]","total":-1,"completed":11,"skipped":63,"failed":0}
[BeforeEach] [sig-node] KubeletManagedEtcHosts
  /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:185
STEP: Creating a kubernetes client
Jul 18 19:55:51.667: INFO: >>> kubeConfig: /root/.kube/config
STEP: Building a namespace api object, basename e2e-kubelet-etc-hosts
STEP: Waiting for a default service account to be provisioned in namespace
... skipping 59 lines ...
• [SLOW TEST:27.558 seconds]
[sig-node] KubeletManagedEtcHosts
/workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/common/node/framework.go:23
  should test kubelet managed /etc/hosts file [LinuxOnly] [NodeConformance] [Conformance]
  /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:630
------------------------------
{"msg":"PASSED [sig-node] KubeletManagedEtcHosts should test kubelet managed /etc/hosts file [LinuxOnly] [NodeConformance] [Conformance]","total":-1,"completed":12,"skipped":63,"failed":0}

S
------------------------------
[BeforeEach] [Testpattern: Dynamic PV (default fs)] volumes
  /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/storage/framework/testsuite.go:51
Jul 18 19:56:19.238: INFO: Only supported for providers [openstack] (not skeleton)
... skipping 211 lines ...
      Only supported for providers [gce gke] (not skeleton)

      /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/storage/drivers/in_tree.go:1301
------------------------------
S
------------------------------
{"msg":"PASSED [sig-storage] CSI Volumes [Driver: csi-hostpath] [Testpattern: Dynamic PV (block volmode)] volumes should store data","total":-1,"completed":10,"skipped":115,"failed":0}
[BeforeEach] [sig-api-machinery] Garbage collector
  /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:185
STEP: Creating a kubernetes client
Jul 18 19:56:16.756: INFO: >>> kubeConfig: /root/.kube/config
STEP: Building a namespace api object, basename gc
STEP: Waiting for a default service account to be provisioned in namespace
... skipping 32 lines ...
  /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:186
Jul 18 19:56:19.455: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready
STEP: Destroying namespace "gc-6234" for this suite.

•
------------------------------
{"msg":"PASSED [sig-api-machinery] Garbage collector should delete RS created by deployment when not orphaning [Conformance]","total":-1,"completed":11,"skipped":115,"failed":0}

S
------------------------------
[BeforeEach] [sig-api-machinery] Watchers
  /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:185
STEP: Creating a kubernetes client
... skipping 14 lines ...
• [SLOW TEST:5.836 seconds]
[sig-api-machinery] Watchers
/workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/apimachinery/framework.go:23
  should receive events on concurrent watches in same order [Conformance]
  /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:630
------------------------------
{"msg":"PASSED [sig-api-machinery] Watchers should receive events on concurrent watches in same order [Conformance]","total":-1,"completed":11,"skipped":45,"failed":0}

SS
------------------------------
[BeforeEach] [Testpattern: Dynamic PV (block volmode)(allowExpansion)] volume-expand
  /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/storage/framework/testsuite.go:51
Jul 18 19:56:19.806: INFO: Driver local doesn't support DynamicPV -- skipping
... skipping 30071 lines ...






efulSet ss2 successful\"\nE0718 20:00:08.506342       1 tokens_controller.go:262] error synchronizing serviceaccount provisioning-8696/default: serviceaccounts \"default\" not found\nI0718 20:00:09.787241       1 namespace_controller.go:185] Namespace has been deleted kubectl-3445\nI0718 20:00:10.428247       1 replica_set.go:559] \"Too few replicas\" replicaSet=\"deployment-672/test-rollover-deployment-78bc8b888c\" need=1 creating=1\nI0718 20:00:10.428391       1 event.go:291] \"Event occurred\" object=\"deployment-672/test-rollover-deployment\" kind=\"Deployment\" apiVersion=\"apps/v1\" type=\"Normal\" reason=\"ScalingReplicaSet\" message=\"Scaled up replica set test-rollover-deployment-78bc8b888c to 1\"\nI0718 20:00:10.436039       1 event.go:291] \"Event occurred\" object=\"deployment-672/test-rollover-deployment-78bc8b888c\" kind=\"ReplicaSet\" apiVersion=\"apps/v1\" type=\"Normal\" reason=\"SuccessfulCreate\" message=\"Created pod: test-rollover-deployment-78bc8b888c-nb9tb\"\nI0718 20:00:10.444891       1 deployment_controller.go:490] \"Error syncing deployment\" deployment=\"deployment-672/test-rollover-deployment\" err=\"Operation cannot be fulfilled on deployments.apps \\\"test-rollover-deployment\\\": the object has been modified; please apply your changes to the latest version and try again\"\nI0718 20:00:10.462029       1 deployment_controller.go:490] \"Error syncing deployment\" deployment=\"deployment-672/test-rollover-deployment\" err=\"Operation cannot be fulfilled on deployments.apps \\\"test-rollover-deployment\\\": the object has been modified; please apply your changes to the latest version and try again\"\nE0718 20:00:10.569573       1 reflector.go:138] k8s.io/client-go/metadata/metadatainformer/informer.go:90: Failed to watch *v1.PartialObjectMetadata: failed to list *v1.PartialObjectMetadata: the server could not find the requested resource\nE0718 20:00:10.626839       1 tokens_controller.go:262] error synchronizing serviceaccount provisioning-5299/default: secrets \"default-token-6wbgz\" is forbidden: unable to create new content in namespace provisioning-5299 because it is being terminated\nI0718 20:00:10.696897       1 replica_set.go:595] \"Too many replicas\" replicaSet=\"deployment-672/test-rollover-deployment-78bc8b888c\" need=0 deleting=1\nI0718 20:00:10.696927       1 replica_set.go:223] \"Found related ReplicaSets\" replicaSet=\"deployment-672/test-rollover-deployment-78bc8b888c\" relatedReplicaSets=[test-rollover-deployment-98c5f4599 test-rollover-controller test-rollover-deployment-78bc8b888c]\nI0718 20:00:10.696975       1 controller_utils.go:602] \"Deleting pod\" controller=\"test-rollover-deployment-78bc8b888c\" pod=\"deployment-672/test-rollover-deployment-78bc8b888c-nb9tb\"\nI0718 20:00:10.697085       1 event.go:291] \"Event occurred\" object=\"deployment-672/test-rollover-deployment\" kind=\"Deployment\" apiVersion=\"apps/v1\" type=\"Normal\" reason=\"ScalingReplicaSet\" message=\"Scaled down replica set test-rollover-deployment-78bc8b888c to 0\"\nI0718 20:00:10.705391       1 deployment_controller.go:490] \"Error syncing deployment\" deployment=\"deployment-672/test-rollover-deployment\" err=\"Operation cannot be fulfilled on deployments.apps \\\"test-rollover-deployment\\\": the object has been modified; please apply your changes to the latest version and try again\"\nI0718 20:00:10.709202       1 event.go:291] \"Event occurred\" object=\"deployment-672/test-rollover-deployment-78bc8b888c\" kind=\"ReplicaSet\" apiVersion=\"apps/v1\" type=\"Normal\" reason=\"SuccessfulDelete\" message=\"Deleted pod: test-rollover-deployment-78bc8b888c-nb9tb\"\nI0718 20:00:10.717070       1 replica_set.go:559] \"Too few replicas\" replicaSet=\"deployment-672/test-rollover-deployment-98c5f4599\" need=1 creating=1\nI0718 20:00:10.718325       1 event.go:291] \"Event occurred\" object=\"deployment-672/test-rollover-deployment\" kind=\"Deployment\" apiVersion=\"apps/v1\" type=\"Normal\" reason=\"ScalingReplicaSet\" message=\"Scaled up replica set test-rollover-deployment-98c5f4599 to 1\"\nI0718 20:00:10.730737       1 event.go:291] \"Event occurred\" object=\"deployment-672/test-rollover-deployment-98c5f4599\" kind=\"ReplicaSet\" apiVersion=\"apps/v1\" type=\"Normal\" reason=\"SuccessfulCreate\" message=\"Created pod: test-rollover-deployment-98c5f4599-tf876\"\nE0718 20:00:10.769499       1 tokens_controller.go:262] error synchronizing serviceaccount conntrack-784/default: secrets \"default-token-t7xbj\" is forbidden: unable to create new content in namespace conntrack-784 because it is being terminated\nI0718 20:00:10.889290       1 namespace_controller.go:185] Namespace has been deleted apply-4061\nI0718 20:00:11.219709       1 graph_builder.go:587] add [v1/Pod, namespace: ephemeral-8120, name: inline-volume-tester-vk42d, uid: 7426787e-7b82-4f79-812b-61d8025ee5ae] to the attemptToDelete, because it's waiting for its dependents to be deleted\nI0718 20:00:11.219850       1 garbagecollector.go:471] \"Processing object\" object=\"ephemeral-8120/inline-volume-tester-vk42d\" objectUID=7426787e-7b82-4f79-812b-61d8025ee5ae kind=\"Pod\" virtual=false\nI0718 20:00:11.236888       1 garbagecollector.go:590] remove DeleteDependents finalizer for item [v1/Pod, namespace: ephemeral-8120, name: inline-volume-tester-vk42d, uid: 7426787e-7b82-4f79-812b-61d8025ee5ae]\nE0718 20:00:11.811016       1 reflector.go:138] k8s.io/client-go/metadata/metadatainformer/informer.go:90: Failed to watch *v1.PartialObjectMetadata: failed to list *v1.PartialObjectMetadata: the server could not find the requested resource\nI0718 20:00:11.931129       1 namespace_controller.go:185] Namespace has been deleted provisioning-2470\nI0718 20:00:13.613359       1 namespace_controller.go:185] Namespace has been deleted provisioning-8696\nI0718 20:00:14.022915       1 pvc_protection_controller.go:291] \"PVC is unused\" PVC=\"provisioning-5077/pvc-s7kzp\"\nI0718 20:00:14.034574       1 pv_controller.go:640] volume \"local-6xbw8\" is released and reclaim policy \"Retain\" will be executed\nI0718 20:00:14.039826       1 pv_controller.go:879] volume \"local-6xbw8\" entered phase \"Released\"\nI0718 20:00:14.068992       1 pv_controller_base.go:505] deletion of claim \"provisioning-5077/pvc-s7kzp\" was already processed\nE0718 20:00:14.232041       1 pv_controller.go:1452] error finding provisioning plugin for claim provisioning-1938/pvc-n5sq9: storageclass.storage.k8s.io \"provisioning-1938\" not found\nI0718 20:00:14.232098       1 event.go:291] \"Event occurred\" object=\"provisioning-1938/pvc-n5sq9\" kind=\"PersistentVolumeClaim\" apiVersion=\"v1\" type=\"Warning\" reason=\"ProvisioningFailed\" message=\"storageclass.storage.k8s.io \\\"provisioning-1938\\\" not found\"\nI0718 20:00:14.272495       1 pv_controller.go:879] volume \"local-5jc7w\" entered phase \"Available\"\nI0718 20:00:14.410230       1 garbagecollector.go:471] \"Processing object\" object=\"ephemeral-2453-2953/csi-hostpath-attacher-4k8jc\" objectUID=97fb8173-9efa-4069-9cc5-eba86a535d93 kind=\"EndpointSlice\" virtual=false\nI0718 20:00:14.421266       1 garbagecollector.go:580] \"Deleting object\" object=\"ephemeral-2453-2953/csi-hostpath-attacher-4k8jc\" objectUID=97fb8173-9efa-4069-9cc5-eba86a535d93 kind=\"EndpointSlice\" propagationPolicy=Background\nI0718 20:00:14.442158       1 replica_set.go:595] \"Too many replicas\" replicaSet=\"deployment-3142/test-deployment-748588b7cd\" need=0 deleting=1\nI0718 20:00:14.442195       1 replica_set.go:223] \"Found related ReplicaSets\" replicaSet=\"deployment-3142/test-deployment-748588b7cd\" relatedReplicaSets=[test-deployment-7b4c744884 test-deployment-748588b7cd test-deployment-85d87c6f4b]\nI0718 20:00:14.442265       1 controller_utils.go:602] \"Deleting pod\" controller=\"test-deployment-748588b7cd\" pod=\"deployment-3142/test-deployment-748588b7cd-vtdnh\"\nI0718 20:00:14.442304       1 event.go:291] \"Event occurred\" object=\"deployment-3142/test-deployment\" kind=\"Deployment\" apiVersion=\"apps/v1\" type=\"Normal\" reason=\"ScalingReplicaSet\" message=\"Scaled down replica set test-deployment-748588b7cd to 0\"\nI0718 20:00:14.455472       1 event.go:291] \"Event occurred\" object=\"deployment-3142/test-deployment-748588b7cd\" kind=\"ReplicaSet\" apiVersion=\"apps/v1\" type=\"Normal\" reason=\"SuccessfulDelete\" message=\"Deleted pod: test-deployment-748588b7cd-vtdnh\"\nI0718 20:00:14.472284       1 garbagecollector.go:471] \"Processing object\" object=\"ephemeral-2453-2953/csi-hostpath-attacher-78556f9d9b\" objectUID=8289f7f8-e8db-4d40-8c8c-0ae7b49e53d7 kind=\"ControllerRevision\" virtual=false\nI0718 20:00:14.472386       1 stateful_set.go:419] StatefulSet has been deleted ephemeral-2453-2953/csi-hostpath-attacher\nI0718 20:00:14.472392       1 garbagecollector.go:471] \"Processing object\" object=\"ephemeral-2453-2953/csi-hostpath-attacher-0\" objectUID=057128e0-3850-4859-ad65-47cd1e021954 kind=\"Pod\" virtual=false\nI0718 20:00:14.478309       1 garbagecollector.go:580] \"Deleting object\" object=\"ephemeral-2453-2953/csi-hostpath-attacher-78556f9d9b\" objectUID=8289f7f8-e8db-4d40-8c8c-0ae7b49e53d7 kind=\"ControllerRevision\" propagationPolicy=Background\nI0718 20:00:14.479181       1 garbagecollector.go:580] \"Deleting object\" object=\"ephemeral-2453-2953/csi-hostpath-attacher-0\" objectUID=057128e0-3850-4859-ad65-47cd1e021954 kind=\"Pod\" propagationPolicy=Background\nI0718 20:00:14.544631       1 garbagecollector.go:471] \"Processing object\" object=\"ephemeral-2453-2953/csi-hostpathplugin-66qxd\" objectUID=962c5d12-3418-493a-b5fe-a4883c6c285a kind=\"EndpointSlice\" virtual=false\nI0718 20:00:14.547311       1 garbagecollector.go:580] \"Deleting object\" object=\"ephemeral-2453-2953/csi-hostpathplugin-66qxd\" objectUID=962c5d12-3418-493a-b5fe-a4883c6c285a kind=\"EndpointSlice\" propagationPolicy=Background\nI0718 20:00:14.594967       1 stateful_set.go:419] StatefulSet has been deleted ephemeral-2453-2953/csi-hostpathplugin\nI0718 20:00:14.594970       1 garbagecollector.go:471] \"Processing object\" object=\"ephemeral-2453-2953/csi-hostpathplugin-668c46cbcd\" objectUID=bdfc5a40-ff97-46e4-a922-c186702144ee kind=\"ControllerRevision\" virtual=false\nI0718 20:00:14.595009       1 garbagecollector.go:471] \"Processing object\" object=\"ephemeral-2453-2953/csi-hostpathplugin-0\" objectUID=c2de8372-21dd-42a4-bf49-ecea7162e76a kind=\"Pod\" virtual=false\nI0718 20:00:14.597469       1 garbagecollector.go:580] \"Deleting object\" object=\"ephemeral-2453-2953/csi-hostpathplugin-0\" objectUID=c2de8372-21dd-42a4-bf49-ecea7162e76a kind=\"Pod\" propagationPolicy=Background\nI0718 20:00:14.598274       1 garbagecollector.go:580] \"Deleting object\" object=\"ephemeral-2453-2953/csi-hostpathplugin-668c46cbcd\" objectUID=bdfc5a40-ff97-46e4-a922-c186702144ee kind=\"ControllerRevision\" propagationPolicy=Background\nI0718 20:00:14.628988       1 garbagecollector.go:471] \"Processing object\" object=\"ephemeral-2453-2953/csi-hostpath-provisioner-q4hnd\" objectUID=da8b294e-8c46-4e49-897a-55ffa874f6dd kind=\"EndpointSlice\" virtual=false\nI0718 20:00:14.631522       1 garbagecollector.go:580] \"Deleting object\" object=\"ephemeral-2453-2953/csi-hostpath-provisioner-q4hnd\" objectUID=da8b294e-8c46-4e49-897a-55ffa874f6dd kind=\"EndpointSlice\" propagationPolicy=Background\nI0718 20:00:14.638212       1 garbagecollector.go:471] \"Processing object\" object=\"deployment-3142/test-deployment-7b4c744884\" objectUID=51a1cc26-6dd4-4de2-9232-43d558bf59fe kind=\"ReplicaSet\" virtual=false\nI0718 20:00:14.638238       1 deployment_controller.go:583] \"Deployment has been deleted\" deployment=\"deployment-3142/test-deployment\"\nI0718 20:00:14.638241       1 garbagecollector.go:471] \"Processing object\" object=\"deployment-3142/test-deployment-748588b7cd\" objectUID=90a037df-bc48-41b4-ab9e-e4d6ac9a430b kind=\"ReplicaSet\" virtual=false\nI0718 20:00:14.638253       1 garbagecollector.go:471] \"Processing object\" object=\"deployment-3142/test-deployment-85d87c6f4b\" objectUID=2db1027d-89d2-40ec-8f36-facc7338e147 kind=\"ReplicaSet\" virtual=false\nI0718 20:00:14.643887       1 garbagecollector.go:580] \"Deleting object\" object=\"deployment-3142/test-deployment-748588b7cd\" objectUID=90a037df-bc48-41b4-ab9e-e4d6ac9a430b kind=\"ReplicaSet\" propagationPolicy=Background\nI0718 20:00:14.643939       1 garbagecollector.go:580] \"Deleting object\" object=\"deployment-3142/test-deployment-85d87c6f4b\" objectUID=2db1027d-89d2-40ec-8f36-facc7338e147 kind=\"ReplicaSet\" propagationPolicy=Background\nI0718 20:00:14.644017       1 garbagecollector.go:580] \"Deleting object\" object=\"deployment-3142/test-deployment-7b4c744884\" objectUID=51a1cc26-6dd4-4de2-9232-43d558bf59fe kind=\"ReplicaSet\" propagationPolicy=Background\nI0718 20:00:14.650634       1 garbagecollector.go:471] \"Processing object\" object=\"deployment-3142/test-deployment-85d87c6f4b-p8wh2\" objectUID=39685246-e062-4fba-af07-31c17c369119 kind=\"Pod\" virtual=false\nI0718 20:00:14.650661       1 garbagecollector.go:471] \"Processing object\" object=\"deployment-3142/test-deployment-85d87c6f4b-krkh2\" objectUID=ea0df1ca-097f-4d4b-8a69-f7306886b388 kind=\"Pod\" virtual=false\nI0718 20:00:14.658446       1 garbagecollector.go:580] \"Deleting object\" object=\"deployment-3142/test-deployment-85d87c6f4b-p8wh2\" objectUID=39685246-e062-4fba-af07-31c17c369119 kind=\"Pod\" propagationPolicy=Background\nI0718 20:00:14.658457       1 garbagecollector.go:580] \"Deleting object\" object=\"deployment-3142/test-deployment-85d87c6f4b-krkh2\" objectUID=ea0df1ca-097f-4d4b-8a69-f7306886b388 kind=\"Pod\" propagationPolicy=Background\nI0718 20:00:14.683582       1 stateful_set.go:419] StatefulSet has been deleted ephemeral-2453-2953/csi-hostpath-provisioner\nI0718 20:00:14.683605       1 garbagecollector.go:471] \"Processing object\" object=\"ephemeral-2453-2953/csi-hostpath-provisioner-6d8f755c74\" objectUID=ab128098-7d6c-479d-adc8-0fe40558f261 kind=\"ControllerRevision\" virtual=false\nI0718 20:00:14.683609       1 garbagecollector.go:471] \"Processing object\" object=\"ephemeral-2453-2953/csi-hostpath-provisioner-0\" objectUID=c6b5be91-219d-4264-bdaa-b149638bdeb4 kind=\"Pod\" virtual=false\nI0718 20:00:14.686366       1 garbagecollector.go:580] \"Deleting object\" object=\"ephemeral-2453-2953/csi-hostpath-provisioner-0\" objectUID=c6b5be91-219d-4264-bdaa-b149638bdeb4 kind=\"Pod\" propagationPolicy=Background\nI0718 20:00:14.687311       1 garbagecollector.go:580] \"Deleting object\" object=\"ephemeral-2453-2953/csi-hostpath-provisioner-6d8f755c74\" objectUID=ab128098-7d6c-479d-adc8-0fe40558f261 kind=\"ControllerRevision\" propagationPolicy=Background\nI0718 20:00:14.725653       1 garbagecollector.go:471] \"Processing object\" object=\"ephemeral-2453-2953/csi-hostpath-resizer-gh628\" objectUID=75171f19-1b45-4775-99d5-6d675417d9a1 kind=\"EndpointSlice\" virtual=false\nI0718 20:00:14.729812       1 garbagecollector.go:580] \"Deleting object\" object=\"ephemeral-2453-2953/csi-hostpath-resizer-gh628\" objectUID=75171f19-1b45-4775-99d5-6d675417d9a1 kind=\"EndpointSlice\" propagationPolicy=Background\nI0718 20:00:14.761741       1 namespace_controller.go:185] Namespace has been deleted downward-api-4806\nI0718 20:00:14.789547       1 stateful_set.go:419] StatefulSet has been deleted ephemeral-2453-2953/csi-hostpath-resizer\nI0718 20:00:14.789555       1 garbagecollector.go:471] \"Processing object\" object=\"ephemeral-2453-2953/csi-hostpath-resizer-79f5c94dbc\" objectUID=490ff52f-c599-4584-a585-24c8da18231c kind=\"ControllerRevision\" virtual=false\nI0718 20:00:14.789592       1 garbagecollector.go:471] \"Processing object\" object=\"ephemeral-2453-2953/csi-hostpath-resizer-0\" objectUID=3dcde969-19d2-4bce-b658-31d1e5b955ed kind=\"Pod\" virtual=false\nI0718 20:00:14.792211       1 garbagecollector.go:580] \"Deleting object\" object=\"ephemeral-2453-2953/csi-hostpath-resizer-0\" objectUID=3dcde969-19d2-4bce-b658-31d1e5b955ed kind=\"Pod\" propagationPolicy=Background\nI0718 20:00:14.792820       1 garbagecollector.go:580] \"Deleting object\" object=\"ephemeral-2453-2953/csi-hostpath-resizer-79f5c94dbc\" objectUID=490ff52f-c599-4584-a585-24c8da18231c kind=\"ControllerRevision\" propagationPolicy=Background\nI0718 20:00:14.836843       1 garbagecollector.go:471] \"Processing object\" object=\"ephemeral-2453-2953/csi-hostpath-snapshotter-ncgdk\" objectUID=bb04dac7-79d2-4615-a094-f449e16a7ba9 kind=\"EndpointSlice\" virtual=false\nI0718 20:00:14.839816       1 garbagecollector.go:580] \"Deleting object\" object=\"ephemeral-2453-2953/csi-hostpath-snapshotter-ncgdk\" objectUID=bb04dac7-79d2-4615-a094-f449e16a7ba9 kind=\"EndpointSlice\" propagationPolicy=Background\nI0718 20:00:14.898795       1 garbagecollector.go:471] \"Processing object\" object=\"ephemeral-2453-2953/csi-hostpath-snapshotter-6c898ddb7\" objectUID=d757113c-d4a5-42bc-bbae-2f8cc317a14d kind=\"ControllerRevision\" virtual=false\nI0718 20:00:14.898827       1 stateful_set.go:419] StatefulSet has been deleted ephemeral-2453-2953/csi-hostpath-snapshotter\nI0718 20:00:14.898894       1 garbagecollector.go:471] \"Processing object\" object=\"ephemeral-2453-2953/csi-hostpath-snapshotter-0\" objectUID=0e13bd65-ecb5-4ce0-a090-e3253c693fd7 kind=\"Pod\" virtual=false\nI0718 20:00:14.901443       1 garbagecollector.go:580] \"Deleting object\" object=\"ephemeral-2453-2953/csi-hostpath-snapshotter-6c898ddb7\" objectUID=d757113c-d4a5-42bc-bbae-2f8cc317a14d kind=\"ControllerRevision\" propagationPolicy=Background\nI0718 20:00:14.902495       1 garbagecollector.go:580] \"Deleting object\" object=\"ephemeral-2453-2953/csi-hostpath-snapshotter-0\" objectUID=0e13bd65-ecb5-4ce0-a090-e3253c693fd7 kind=\"Pod\" propagationPolicy=Background\nE0718 20:00:15.967713       1 pv_controller.go:1452] error finding provisioning plugin for claim provisioning-8250/pvc-ssjhx: storageclass.storage.k8s.io \"provisioning-8250\" not found\nI0718 20:00:15.967811       1 event.go:291] \"Event occurred\" object=\"provisioning-8250/pvc-ssjhx\" kind=\"PersistentVolumeClaim\" apiVersion=\"v1\" type=\"Warning\" reason=\"ProvisioningFailed\" message=\"storageclass.storage.k8s.io \\\"provisioning-8250\\\" not found\"\nI0718 20:00:16.008284       1 pv_controller.go:879] volume \"local-fzxpt\" entered phase \"Available\"\nI0718 20:00:16.217456       1 replica_set.go:449] ReplicaSet \"test-rollover-deployment-98c5f4599\" will be enqueued after 10s for availability check\nI0718 20:00:16.237265       1 namespace_controller.go:185] Namespace has been deleted provisioning-5299\nI0718 20:00:16.306569       1 namespace_controller.go:185] Namespace has been deleted conntrack-784\nI0718 20:00:16.667264       1 event.go:291] \"Event occurred\" object=\"job-2919/backofflimit\" kind=\"Job\" apiVersion=\"batch/v1\" type=\"Normal\" reason=\"SuccessfulCreate\" message=\"Created pod: backofflimit-jmzv7\"\nI0718 20:00:17.846803       1 namespace_controller.go:185] Namespace has been deleted ephemeral-2453\nI0718 20:00:18.240300       1 garbagecollector.go:471] \"Processing object\" object=\"statefulset-8233/ss2-677d6db895\" objectUID=67d0a4f5-b9e7-4e52-bce3-a9036f58e65e kind=\"ControllerRevision\" virtual=false\nI0718 20:00:18.240359       1 stateful_set.go:419] StatefulSet has been deleted statefulset-8233/ss2\nI0718 20:00:18.240394       1 garbagecollector.go:471] \"Processing object\" object=\"statefulset-8233/ss2-5bbbc9fc94\" objectUID=61dbfa07-6609-4489-8574-53e5e70f1d9c kind=\"ControllerRevision\" virtual=false\nI0718 20:00:18.243391       1 garbagecollector.go:580] \"Deleting object\" object=\"statefulset-8233/ss2-677d6db895\" objectUID=67d0a4f5-b9e7-4e52-bce3-a9036f58e65e kind=\"ControllerRevision\" propagationPolicy=Background\nI0718 20:00:18.244106       1 garbagecollector.go:580] \"Deleting object\" object=\"statefulset-8233/ss2-5bbbc9fc94\" objectUID=61dbfa07-6609-4489-8574-53e5e70f1d9c kind=\"ControllerRevision\" propagationPolicy=Background\nE0718 20:00:18.634547       1 pv_controller.go:1452] error finding provisioning plugin for claim ephemeral-5827/inline-volume-lp8v6-my-volume: storageclass.storage.k8s.io \"no-such-storage-class\" not found\nI0718 20:00:18.634651       1 event.go:291] \"Event occurred\" object=\"ephemeral-5827/inline-volume-lp8v6-my-volume\" kind=\"PersistentVolumeClaim\" apiVersion=\"v1\" type=\"Warning\" reason=\"ProvisioningFailed\" message=\"storageclass.storage.k8s.io \\\"no-such-storage-class\\\" not found\"\nI0718 20:00:18.720762       1 graph_builder.go:587] add [v1/Pod, namespace: ephemeral-5827, name: inline-volume-lp8v6, uid: 6e337d3f-570d-4d66-b93a-c9711ac62d44] to the attemptToDelete, because it's waiting for its dependents to be deleted\nI0718 20:00:18.720812       1 garbagecollector.go:471] \"Processing object\" object=\"ephemeral-5827/inline-volume-lp8v6-my-volume\" objectUID=af25ae3d-5fea-4d4e-9a60-ad9bfdf08eb1 kind=\"PersistentVolumeClaim\" virtual=false\nI0718 20:00:18.720831       1 garbagecollector.go:471] \"Processing object\" object=\"ephemeral-5827/inline-volume-lp8v6\" objectUID=6e337d3f-570d-4d66-b93a-c9711ac62d44 kind=\"Pod\" virtual=false\nI0718 20:00:18.724115       1 garbagecollector.go:595] adding [v1/PersistentVolumeClaim, namespace: ephemeral-5827, name: inline-volume-lp8v6-my-volume, uid: af25ae3d-5fea-4d4e-9a60-ad9bfdf08eb1] to attemptToDelete, because its owner [v1/Pod, namespace: ephemeral-5827, name: inline-volume-lp8v6, uid: 6e337d3f-570d-4d66-b93a-c9711ac62d44] is deletingDependents\nI0718 20:00:18.725931       1 garbagecollector.go:580] \"Deleting object\" object=\"ephemeral-5827/inline-volume-lp8v6-my-volume\" objectUID=af25ae3d-5fea-4d4e-9a60-ad9bfdf08eb1 kind=\"PersistentVolumeClaim\" propagationPolicy=Background\nI0718 20:00:18.732608       1 garbagecollector.go:471] \"Processing object\" object=\"ephemeral-5827/inline-volume-lp8v6-my-volume\" objectUID=af25ae3d-5fea-4d4e-9a60-ad9bfdf08eb1 kind=\"PersistentVolumeClaim\" virtual=false\nE0718 20:00:18.732681       1 pv_controller.go:1452] error finding provisioning plugin for claim ephemeral-5827/inline-volume-lp8v6-my-volume: storageclass.storage.k8s.io \"no-such-storage-class\" not found\nI0718 20:00:18.732740       1 event.go:291] \"Event occurred\" object=\"ephemeral-5827/inline-volume-lp8v6-my-volume\" kind=\"PersistentVolumeClaim\" apiVersion=\"v1\" type=\"Warning\" reason=\"ProvisioningFailed\" message=\"storageclass.storage.k8s.io \\\"no-such-storage-class\\\" not found\"\nI0718 20:00:18.735280       1 pvc_protection_controller.go:291] \"PVC is unused\" PVC=\"ephemeral-5827/inline-volume-lp8v6-my-volume\"\nI0718 20:00:18.742385       1 garbagecollector.go:471] \"Processing object\" object=\"ephemeral-5827/inline-volume-lp8v6\" objectUID=6e337d3f-570d-4d66-b93a-c9711ac62d44 kind=\"Pod\" virtual=false\nI0718 20:00:18.744610       1 garbagecollector.go:590] remove DeleteDependents finalizer for item [v1/Pod, namespace: ephemeral-5827, name: inline-volume-lp8v6, uid: 6e337d3f-570d-4d66-b93a-c9711ac62d44]\nE0718 20:00:19.057053       1 reflector.go:138] k8s.io/client-go/metadata/metadatainformer/informer.go:90: Failed to watch *v1.PartialObjectMetadata: failed to list *v1.PartialObjectMetadata: the server could not find the requested resource\nE0718 20:00:20.339902       1 tokens_controller.go:262] error synchronizing serviceaccount ephemeral-2453-2953/default: secrets \"default-token-kbx4h\" is forbidden: unable to create new content in namespace ephemeral-2453-2953 because it is being terminated\nI0718 20:00:21.564640       1 pv_controller.go:930] claim \"provisioning-1938/pvc-n5sq9\" bound to volume \"local-5jc7w\"\nI0718 20:00:21.575130       1 pv_controller.go:879] volume \"local-5jc7w\" entered phase \"Bound\"\nI0718 20:00:21.575156       1 pv_controller.go:982] volume \"local-5jc7w\" bound to claim \"provisioning-1938/pvc-n5sq9\"\nI0718 20:00:21.588250       1 pv_controller.go:823] claim \"provisioning-1938/pvc-n5sq9\" entered phase \"Bound\"\nI0718 20:00:21.588413       1 pv_controller.go:930] claim \"provisioning-8250/pvc-ssjhx\" bound to volume \"local-fzxpt\"\nI0718 20:00:21.601072       1 pv_controller.go:879] volume \"local-fzxpt\" entered phase \"Bound\"\nI0718 20:00:21.601098       1 pv_controller.go:982] volume \"local-fzxpt\" bound to claim \"provisioning-8250/pvc-ssjhx\"\nI0718 20:00:21.614964       1 pv_controller.go:823] claim \"provisioning-8250/pvc-ssjhx\" entered phase \"Bound\"\nI0718 20:00:21.822006       1 event.go:291] \"Event occurred\" object=\"ephemeral-5827-4412/csi-hostpath-attacher\" kind=\"StatefulSet\" apiVersion=\"apps/v1\" type=\"Normal\" reason=\"SuccessfulCreate\" message=\"create Pod csi-hostpath-attacher-0 in StatefulSet csi-hostpath-attacher successful\"\nI0718 20:00:21.941670       1 event.go:291] \"Event occurred\" object=\"ephemeral-5827-4412/csi-hostpathplugin\" kind=\"StatefulSet\" apiVersion=\"apps/v1\" type=\"Normal\" reason=\"SuccessfulCreate\" message=\"create Pod csi-hostpathplugin-0 in StatefulSet csi-hostpathplugin successful\"\nI0718 20:00:22.019644       1 event.go:291] \"Event occurred\" object=\"ephemeral-5827-4412/csi-hostpath-provisioner\" kind=\"StatefulSet\" apiVersion=\"apps/v1\" type=\"Normal\" reason=\"SuccessfulCreate\" message=\"create Pod csi-hostpath-provisioner-0 in StatefulSet csi-hostpath-provisioner successful\"\nI0718 20:00:22.102662       1 event.go:291] \"Event occurred\" object=\"ephemeral-5827-4412/csi-hostpath-resizer\" kind=\"StatefulSet\" apiVersion=\"apps/v1\" type=\"Normal\" reason=\"SuccessfulCreate\" message=\"create Pod csi-hostpath-resizer-0 in StatefulSet csi-hostpath-resizer successful\"\nI0718 20:00:22.180874       1 event.go:291] \"Event occurred\" object=\"ephemeral-5827-4412/csi-hostpath-snapshotter\" kind=\"StatefulSet\" apiVersion=\"apps/v1\" type=\"Normal\" reason=\"SuccessfulCreate\" message=\"create Pod csi-hostpath-snapshotter-0 in StatefulSet csi-hostpath-snapshotter successful\"\nI0718 20:00:22.526560       1 event.go:291] \"Event occurred\" object=\"ephemeral-5827/inline-volume-tester-knktf-my-volume-0\" kind=\"PersistentVolumeClaim\" apiVersion=\"v1\" type=\"Normal\" reason=\"WaitForPodScheduled\" message=\"waiting for pod inline-volume-tester-knktf to be scheduled\"\nI0718 20:00:23.004637       1 namespace_controller.go:185] Namespace has been deleted svcaccounts-2162\nI0718 20:00:23.763759       1 event.go:291] \"Event occurred\" object=\"ephemeral-5827/inline-volume-tester-knktf-my-volume-0\" kind=\"PersistentVolumeClaim\" apiVersion=\"v1\" type=\"Normal\" reason=\"ExternalProvisioning\" message=\"waiting for a volume to be created, either by external provisioner \\\"csi-hostpath-ephemeral-5827\\\" or manually created by system administrator\"\nI0718 20:00:23.763966       1 event.go:291] \"Event occurred\" object=\"ephemeral-5827/inline-volume-tester-knktf-my-volume-0\" kind=\"PersistentVolumeClaim\" apiVersion=\"v1\" type=\"Normal\" reason=\"ExternalProvisioning\" message=\"waiting for a volume to be created, either by external provisioner \\\"csi-hostpath-ephemeral-5827\\\" or manually created by system administrator\"\nI0718 20:00:24.702924       1 namespace_controller.go:185] Namespace has been deleted provisioning-5077\nE0718 20:00:24.812424       1 reflector.go:138] k8s.io/client-go/metadata/metadatainformer/informer.go:90: Failed to watch *v1.PartialObjectMetadata: failed to list *v1.PartialObjectMetadata: the server could not find the requested resource\nI0718 20:00:25.331056       1 namespace_controller.go:185] Namespace has been deleted deployment-3142\nI0718 20:00:26.247967       1 replica_set.go:595] \"Too many replicas\" replicaSet=\"deployment-672/test-rollover-controller\" need=0 deleting=1\nI0718 20:00:26.248000       1 replica_set.go:223] \"Found related ReplicaSets\" replicaSet=\"deployment-672/test-rollover-controller\" relatedReplicaSets=[test-rollover-deployment-98c5f4599 test-rollover-controller test-rollover-deployment-78bc8b888c]\nI0718 20:00:26.248045       1 controller_utils.go:602] \"Deleting pod\" controller=\"test-rollover-controller\" pod=\"deployment-672/test-rollover-controller-dvsdc\"\nI0718 20:00:26.248086       1 event.go:291] \"Event occurred\" object=\"deployment-672/test-rollover-deployment\" kind=\"Deployment\" apiVersion=\"apps/v1\" type=\"Normal\" reason=\"ScalingReplicaSet\" message=\"Scaled down replica set test-rollover-controller to 0\"\nI0718 20:00:26.266743       1 event.go:291] \"Event occurred\" object=\"deployment-672/test-rollover-controller\" kind=\"ReplicaSet\" apiVersion=\"apps/v1\" type=\"Normal\" reason=\"SuccessfulDelete\" message=\"Deleted pod: test-rollover-controller-dvsdc\"\nI0718 20:00:26.673936       1 namespace_controller.go:185] Namespace has been deleted downward-api-1949\nI0718 20:00:27.155662       1 event.go:291] \"Event occurred\" object=\"job-2919/backofflimit\" kind=\"Job\" apiVersion=\"batch/v1\" type=\"Normal\" reason=\"SuccessfulCreate\" message=\"Created pod: backofflimit-z2c2h\"\nE0718 20:00:27.167317       1 job_controller.go:404] Error syncing job: failed pod(s) detected for job key \"job-2919/backofflimit\"\nE0718 20:00:27.469115       1 reflector.go:138] k8s.io/client-go/metadata/metadatainformer/informer.go:90: Failed to watch *v1.PartialObjectMetadata: failed to list *v1.PartialObjectMetadata: the server could not find the requested resource\nW0718 20:00:27.611351       1 utils.go:265] Service services-7615/service-headless-toggled using reserved endpoint slices label, skipping label service.kubernetes.io/headless: \nI0718 20:00:27.694638       1 namespace_controller.go:185] Namespace has been deleted emptydir-8493\nI0718 20:00:27.830938       1 pvc_protection_controller.go:303] \"Pod uses PVC\" pod=\"ephemeral-6983/inline-volume-tester-dlmx4\" PVC=\"ephemeral-6983/inline-volume-tester-dlmx4-my-volume-0\"\nI0718 20:00:27.830961       1 pvc_protection_controller.go:181] \"Keeping PVC because it is being used\" PVC=\"ephemeral-6983/inline-volume-tester-dlmx4-my-volume-0\"\nI0718 20:00:27.932517       1 reconciler.go:219] attacherDetacher.DetachVolume started for volume \"pvc-6b8beac3-18eb-46bc-8f14-12e8aa90499c\" (UniqueName: \"kubernetes.io/csi/csi-hostpath-ephemeral-6983^9d30a0b3-e802-11eb-ad70-ba528ca6a70f\") on node \"10.116.0.5\" \nI0718 20:00:27.937081       1 operation_generator.go:1483] Verified volume is safe to detach for volume \"pvc-6b8beac3-18eb-46bc-8f14-12e8aa90499c\" (UniqueName: \"kubernetes.io/csi/csi-hostpath-ephemeral-6983^9d30a0b3-e802-11eb-ad70-ba528ca6a70f\") on node \"10.116.0.5\" \nI0718 20:00:28.037123       1 replica_set.go:559] \"Too few replicas\" replicaSet=\"disruption-5763/rs\" need=10 creating=10\nI0718 20:00:28.044875       1 event.go:291] \"Event occurred\" object=\"disruption-5763/rs\" kind=\"ReplicaSet\" apiVersion=\"apps/v1\" type=\"Normal\" reason=\"SuccessfulCreate\" message=\"Created pod: rs-ppfgl\"\nI0718 20:00:28.058419       1 event.go:291] \"Event occurred\" object=\"disruption-5763/rs\" kind=\"ReplicaSet\" apiVersion=\"apps/v1\" type=\"Normal\" reason=\"SuccessfulCreate\" message=\"Created pod: rs-7rvh6\"\nI0718 20:00:28.064533       1 event.go:291] \"Event occurred\" object=\"disruption-5763/rs\" kind=\"ReplicaSet\" apiVersion=\"apps/v1\" type=\"Normal\" reason=\"SuccessfulCreate\" message=\"Created pod: rs-sgqwg\"\nI0718 20:00:28.078058       1 event.go:291] \"Event occurred\" object=\"disruption-5763/rs\" kind=\"ReplicaSet\" apiVersion=\"apps/v1\" type=\"Normal\" reason=\"SuccessfulCreate\" message=\"Created pod: rs-7pflq\"\nI0718 20:00:28.085686       1 event.go:291] \"Event occurred\" object=\"disruption-5763/rs\" kind=\"ReplicaSet\" apiVersion=\"apps/v1\" type=\"Normal\" reason=\"SuccessfulCreate\" message=\"Created pod: rs-hbskc\"\nI0718 20:00:28.085707       1 event.go:291] \"Event occurred\" object=\"disruption-5763/rs\" kind=\"ReplicaSet\" apiVersion=\"apps/v1\" type=\"Normal\" reason=\"SuccessfulCreate\" message=\"Created pod: rs-w2cw5\"\nI0718 20:00:28.085719       1 event.go:291] \"Event occurred\" object=\"disruption-5763/rs\" kind=\"ReplicaSet\" apiVersion=\"apps/v1\" type=\"Normal\" reason=\"SuccessfulCreate\" message=\"Created pod: rs-67rmb\"\nI0718 20:00:28.097020       1 event.go:291] \"Event occurred\" object=\"disruption-5763/rs\" kind=\"ReplicaSet\" apiVersion=\"apps/v1\" type=\"Normal\" reason=\"SuccessfulCreate\" message=\"Created pod: rs-2hk69\"\nI0718 20:00:28.097158       1 event.go:291] \"Event occurred\" object=\"disruption-5763/rs\" kind=\"ReplicaSet\" apiVersion=\"apps/v1\" type=\"Normal\" reason=\"SuccessfulCreate\" message=\"Created pod: rs-qm2bg\"\nI0718 20:00:28.097609       1 event.go:291] \"Event occurred\" object=\"disruption-5763/rs\" kind=\"ReplicaSet\" apiVersion=\"apps/v1\" type=\"Normal\" reason=\"SuccessfulCreate\" message=\"Created pod: rs-62n2z\"\nE0718 20:00:28.464181       1 reflector.go:138] k8s.io/client-go/metadata/metadatainformer/informer.go:90: Failed to watch *v1.PartialObjectMetadata: failed to list *v1.PartialObjectMetadata: the server could not find the requested resource\nI0718 20:00:28.512734       1 operation_generator.go:483] DetachVolume.Detach succeeded for volume \"pvc-6b8beac3-18eb-46bc-8f14-12e8aa90499c\" (UniqueName: \"kubernetes.io/csi/csi-hostpath-ephemeral-6983^9d30a0b3-e802-11eb-ad70-ba528ca6a70f\") on node \"10.116.0.5\" \nI0718 20:00:28.638344       1 pvc_protection_controller.go:291] \"PVC is unused\" PVC=\"ephemeral-6983/inline-volume-tester-dlmx4-my-volume-0\"\nI0718 20:00:28.638649       1 event.go:291] \"Event occurred\" object=\"csi-mock-volumes-5970-9198/csi-mockplugin\" kind=\"StatefulSet\" apiVersion=\"apps/v1\" type=\"Normal\" reason=\"SuccessfulCreate\" message=\"create Pod csi-mockplugin-0 in StatefulSet csi-mockplugin successful\"\nI0718 20:00:28.653084       1 garbagecollector.go:471] \"Processing object\" object=\"ephemeral-6983/inline-volume-tester-dlmx4\" objectUID=dc126a0d-2fdc-4b7e-a216-bc9c6f912785 kind=\"Pod\" virtual=false\nI0718 20:00:28.655509       1 garbagecollector.go:590] remove DeleteDependents finalizer for item [v1/Pod, namespace: ephemeral-6983, name: inline-volume-tester-dlmx4, uid: dc126a0d-2fdc-4b7e-a216-bc9c6f912785]\nI0718 20:00:28.656136       1 pv_controller.go:640] volume \"pvc-6b8beac3-18eb-46bc-8f14-12e8aa90499c\" is released and reclaim policy \"Delete\" will be executed\nI0718 20:00:28.662128       1 pv_controller.go:879] volume \"pvc-6b8beac3-18eb-46bc-8f14-12e8aa90499c\" entered phase \"Released\"\nI0718 20:00:28.664332       1 pv_controller.go:1341] isVolumeReleased[pvc-6b8beac3-18eb-46bc-8f14-12e8aa90499c]: volume is released\nI0718 20:00:28.688558       1 pv_controller_base.go:505] deletion of claim \"ephemeral-6983/inline-volume-tester-dlmx4-my-volume-0\" was already processed\nI0718 20:00:28.808230       1 pv_controller.go:879] volume \"pvc-dc473523-d38a-4923-8439-32b2189b8860\" entered phase \"Bound\"\nI0718 20:00:28.808260       1 pv_controller.go:982] volume \"pvc-dc473523-d38a-4923-8439-32b2189b8860\" bound to claim \"ephemeral-5827/inline-volume-tester-knktf-my-volume-0\"\nI0718 20:00:28.825465       1 pv_controller.go:823] claim \"ephemeral-5827/inline-volume-tester-knktf-my-volume-0\" entered phase \"Bound\"\nI0718 20:00:28.918736       1 namespace_controller.go:185] Namespace has been deleted statefulset-8233\nI0718 20:00:29.841246       1 reconciler.go:295] attacherDetacher.AttachVolume started for volume \"pvc-dc473523-d38a-4923-8439-32b2189b8860\" (UniqueName: \"kubernetes.io/csi/csi-hostpath-ephemeral-5827^ccc704c3-e802-11eb-8233-02fc2257e70c\") from node \"10.116.0.7\" \nI0718 20:00:30.380072       1 operation_generator.go:368] AttachVolume.Attach succeeded for volume \"pvc-dc473523-d38a-4923-8439-32b2189b8860\" (UniqueName: \"kubernetes.io/csi/csi-hostpath-ephemeral-5827^ccc704c3-e802-11eb-8233-02fc2257e70c\") from node \"10.116.0.7\" \nI0718 20:00:30.380193       1 event.go:291] \"Event occurred\" object=\"ephemeral-5827/inline-volume-tester-knktf\" kind=\"Pod\" apiVersion=\"v1\" type=\"Normal\" reason=\"SuccessfulAttachVolume\" message=\"AttachVolume.Attach succeeded for volume \\\"pvc-dc473523-d38a-4923-8439-32b2189b8860\\\" \"\nE0718 20:00:32.086626       1 tokens_controller.go:262] error synchronizing serviceaccount init-container-65/default: secrets \"default-token-k8k9z\" is forbidden: unable to create new content in namespace init-container-65 because it is being terminated\nI0718 20:00:32.197620       1 replica_set.go:559] \"Too few replicas\" replicaSet=\"deployment-672/test-rollover-deployment-98c5f4599\" need=1 creating=1\nI0718 20:00:32.251784       1 garbagecollector.go:471] \"Processing object\" object=\"deployment-672/test-rollover-controller\" objectUID=23ae4b51-b3f7-4add-88a3-b6474f46db8e kind=\"ReplicaSet\" virtual=false\nI0718 20:00:32.251815       1 deployment_controller.go:583] \"Deployment has been deleted\" deployment=\"deployment-672/test-rollover-deployment\"\nI0718 20:00:32.251838       1 garbagecollector.go:471] \"Processing object\" object=\"deployment-672/test-rollover-deployment-98c5f4599\" objectUID=4ab7b5e9-5ed8-4286-aea2-d454913c1a86 kind=\"ReplicaSet\" virtual=false\nI0718 20:00:32.251791       1 garbagecollector.go:471] \"Processing object\" object=\"deployment-672/test-rollover-deployment-78bc8b888c\" objectUID=93d3b5b7-119a-4103-91cf-5987d35a7cd8 kind=\"ReplicaSet\" virtual=false\nI0718 20:00:32.255513       1 garbagecollector.go:580] \"Deleting object\" object=\"deployment-672/test-rollover-deployment-98c5f4599\" objectUID=4ab7b5e9-5ed8-4286-aea2-d454913c1a86 kind=\"ReplicaSet\" propagationPolicy=Background\nI0718 20:00:32.255537       1 garbagecollector.go:580] \"Deleting object\" object=\"deployment-672/test-rollover-controller\" objectUID=23ae4b51-b3f7-4add-88a3-b6474f46db8e kind=\"ReplicaSet\" propagationPolicy=Background\nI0718 20:00:32.255516       1 garbagecollector.go:580] \"Deleting object\" object=\"deployment-672/test-rollover-deployment-78bc8b888c\" objectUID=93d3b5b7-119a-4103-91cf-5987d35a7cd8 kind=\"ReplicaSet\" propagationPolicy=Background\nE0718 20:00:32.304037       1 tokens_controller.go:262] error synchronizing serviceaccount provisioning-6645/default: secrets \"default-token-tc4qv\" is forbidden: unable to create new content in namespace provisioning-6645 because it is being terminated\nE0718 20:00:32.437769       1 tokens_controller.go:262] error synchronizing serviceaccount deployment-672/default: secrets \"default-token-2vqxp\" is forbidden: unable to create new content in namespace deployment-672 because it is being terminated\nE0718 20:00:32.552848       1 tokens_controller.go:262] error synchronizing serviceaccount node-problem-detector-379/default: secrets \"default-token-ksq7s\" is forbidden: unable to create new content in namespace node-problem-detector-379 because it is being terminated\nE0718 20:00:33.027769       1 tokens_controller.go:262] error synchronizing serviceaccount pods-1715/default: serviceaccounts \"default\" not found\nE0718 20:00:34.732707       1 reflector.go:138] k8s.io/client-go/metadata/metadatainformer/informer.go:90: Failed to watch *v1.PartialObjectMetadata: failed to list *v1.PartialObjectMetadata: the server could not find the requested resource\nE0718 20:00:35.434734       1 reflector.go:138] k8s.io/client-go/metadata/metadatainformer/informer.go:90: Failed to watch *v1.PartialObjectMetadata: failed to list *v1.PartialObjectMetadata: the server could not find the requested resource\nI0718 20:00:36.859576       1 garbagecollector.go:471] \"Processing object\" object=\"ephemeral-6983-2204/csi-hostpath-attacher-d7t8q\" objectUID=2141c1ca-3b29-4922-907f-2388d7fd1a87 kind=\"EndpointSlice\" virtual=false\nI0718 20:00:36.865570       1 garbagecollector.go:580] \"Deleting object\" object=\"ephemeral-6983-2204/csi-hostpath-attacher-d7t8q\" objectUID=2141c1ca-3b29-4922-907f-2388d7fd1a87 kind=\"EndpointSlice\" propagationPolicy=Background\nI0718 20:00:36.908279       1 stateful_set.go:419] StatefulSet has been deleted ephemeral-6983-2204/csi-hostpath-attacher\nI0718 20:00:36.908280       1 garbagecollector.go:471] \"Processing object\" object=\"ephemeral-6983-2204/csi-hostpath-attacher-75ccdd9c5d\" objectUID=61c2bf56-371e-4d85-b112-c9036daae7f9 kind=\"ControllerRevision\" virtual=false\nI0718 20:00:36.908308       1 garbagecollector.go:471] \"Processing object\" object=\"ephemeral-6983-2204/csi-hostpath-attacher-0\" objectUID=51f75ea8-f366-4812-ba29-5e6709141491 kind=\"Pod\" virtual=false\nI0718 20:00:36.910995       1 garbagecollector.go:580] \"Deleting object\" object=\"ephemeral-6983-2204/csi-hostpath-attacher-0\" objectUID=51f75ea8-f366-4812-ba29-5e6709141491 kind=\"Pod\" propagationPolicy=Background\nI0718 20:00:36.917022       1 garbagecollector.go:580] \"Deleting object\" object=\"ephemeral-6983-2204/csi-hostpath-attacher-75ccdd9c5d\" objectUID=61c2bf56-371e-4d85-b112-c9036daae7f9 kind=\"ControllerRevision\" propagationPolicy=Background\nI0718 20:00:36.990316       1 garbagecollector.go:471] \"Processing object\" object=\"ephemeral-6983-2204/csi-hostpathplugin-zj8t7\" objectUID=e06fe02a-fa85-44f4-bb91-1cdb58d9c284 kind=\"EndpointSlice\" virtual=false\nI0718 20:00:36.995077       1 garbagecollector.go:580] \"Deleting object\" object=\"ephemeral-6983-2204/csi-hostpathplugin-zj8t7\" objectUID=e06fe02a-fa85-44f4-bb91-1cdb58d9c284 kind=\"EndpointSlice\" propagationPolicy=Background\nI0718 20:00:37.038164       1 garbagecollector.go:471] \"Processing object\" object=\"ephemeral-6983-2204/csi-hostpathplugin-5d95c9b6d4\" objectUID=835c46c4-d65f-492e-9064-3d865805b704 kind=\"ControllerRevision\" virtual=false\nI0718 20:00:37.038176       1 stateful_set.go:419] StatefulSet has been deleted ephemeral-6983-2204/csi-hostpathplugin\nI0718 20:00:37.038190       1 garbagecollector.go:471] \"Processing object\" object=\"ephemeral-6983-2204/csi-hostpathplugin-0\" objectUID=b8cfd421-0c10-4d7b-89ff-35dcf63c2069 kind=\"Pod\" virtual=false\nI0718 20:00:37.042089       1 garbagecollector.go:580] \"Deleting object\" object=\"ephemeral-6983-2204/csi-hostpathplugin-5d95c9b6d4\" objectUID=835c46c4-d65f-492e-9064-3d865805b704 kind=\"ControllerRevision\" propagationPolicy=Background\nI0718 20:00:37.043241       1 garbagecollector.go:580] \"Deleting object\" object=\"ephemeral-6983-2204/csi-hostpathplugin-0\" objectUID=b8cfd421-0c10-4d7b-89ff-35dcf63c2069 kind=\"Pod\" propagationPolicy=Background\nI0718 20:00:37.074930       1 garbagecollector.go:471] \"Processing object\" object=\"ephemeral-6983-2204/csi-hostpath-provisioner-btzfl\" objectUID=184e405e-a2d2-4b7b-8163-2a494755fcf6 kind=\"EndpointSlice\" virtual=false\nI0718 20:00:37.077316       1 garbagecollector.go:580] \"Deleting object\" object=\"ephemeral-6983-2204/csi-hostpath-provisioner-btzfl\" objectUID=184e405e-a2d2-4b7b-8163-2a494755fcf6 kind=\"EndpointSlice\" propagationPolicy=Background\nI0718 20:00:37.095297       1 pvc_protection_controller.go:291] \"PVC is unused\" PVC=\"volumemode-6763/pvc-vd4gw\"\nI0718 20:00:37.107323       1 pv_controller.go:640] volume \"local-9q96t\" is released and reclaim policy \"Retain\" will be executed\nI0718 20:00:37.112964       1 pv_controller.go:879] volume \"local-9q96t\" entered phase \"Released\"\nI0718 20:00:37.122983       1 garbagecollector.go:471] \"Processing object\" object=\"ephemeral-6983-2204/csi-hostpath-provisioner-0\" objectUID=323da5b8-4dc8-4c6f-9096-cfd548709da5 kind=\"Pod\" virtual=false\nI0718 20:00:37.122997       1 garbagecollector.go:471] \"Processing object\" object=\"ephemeral-6983-2204/csi-hostpath-provisioner-649d7d54b8\" objectUID=13655d41-86af-4b0e-9455-ee89603f1145 kind=\"ControllerRevision\" virtual=false\nI0718 20:00:37.123005       1 stateful_set.go:419] StatefulSet has been deleted ephemeral-6983-2204/csi-hostpath-provisioner\nI0718 20:00:37.132856       1 garbagecollector.go:580] \"Deleting object\" object=\"ephemeral-6983-2204/csi-hostpath-provisioner-0\" objectUID=323da5b8-4dc8-4c6f-9096-cfd548709da5 kind=\"Pod\" propagationPolicy=Background\nI0718 20:00:37.133770       1 garbagecollector.go:580] \"Deleting object\" object=\"ephemeral-6983-2204/csi-hostpath-provisioner-649d7d54b8\" objectUID=13655d41-86af-4b0e-9455-ee89603f1145 kind=\"ControllerRevision\" propagationPolicy=Background\nI0718 20:00:37.138929       1 pv_controller_base.go:505] deletion of claim \"volumemode-6763/pvc-vd4gw\" was already processed\nI0718 20:00:37.160938       1 garbagecollector.go:471] \"Processing object\" object=\"ephemeral-6983-2204/csi-hostpath-resizer-wkx7w\" objectUID=26def54f-dab3-4433-bab8-96af7dbc4ffd kind=\"EndpointSlice\" virtual=false\nI0718 20:00:37.163105       1 garbagecollector.go:580] \"Deleting object\" object=\"ephemeral-6983-2204/csi-hostpath-resizer-wkx7w\" objectUID=26def54f-dab3-4433-bab8-96af7dbc4ffd kind=\"EndpointSlice\" propagationPolicy=Background\nI0718 20:00:37.210551       1 garbagecollector.go:471] \"Processing object\" object=\"ephemeral-6983-2204/csi-hostpath-resizer-657687f577\" objectUID=3fe32705-1bc2-4f65-b053-b9880f8dfe66 kind=\"ControllerRevision\" virtual=false\nI0718 20:00:37.210581       1 garbagecollector.go:471] \"Processing object\" object=\"ephemeral-6983-2204/csi-hostpath-resizer-0\" objectUID=a3eda00c-c088-46d0-a52f-65c437bd48db kind=\"Pod\" virtual=false\nI0718 20:00:37.210620       1 stateful_set.go:419] StatefulSet has been deleted ephemeral-6983-2204/csi-hostpath-resizer\nI0718 20:00:37.212756       1 garbagecollector.go:580] \"Deleting object\" object=\"ephemeral-6983-2204/csi-hostpath-resizer-657687f577\" objectUID=3fe32705-1bc2-4f65-b053-b9880f8dfe66 kind=\"ControllerRevision\" propagationPolicy=Background\nI0718 20:00:37.213694       1 garbagecollector.go:580] \"Deleting object\" object=\"ephemeral-6983-2204/csi-hostpath-resizer-0\" objectUID=a3eda00c-c088-46d0-a52f-65c437bd48db kind=\"Pod\" propagationPolicy=Background\nI0718 20:00:37.245678       1 garbagecollector.go:471] \"Processing object\" object=\"ephemeral-6983-2204/csi-hostpath-snapshotter-mx6j2\" objectUID=299f43f7-a5f2-42bd-98b1-14e52ae1d42d kind=\"EndpointSlice\" virtual=false\nI0718 20:00:37.247739       1 garbagecollector.go:580] \"Deleting object\" object=\"ephemeral-6983-2204/csi-hostpath-snapshotter-mx6j2\" objectUID=299f43f7-a5f2-42bd-98b1-14e52ae1d42d kind=\"EndpointSlice\" propagationPolicy=Background\nI0718 20:00:37.295507       1 garbagecollector.go:471] \"Processing object\" object=\"ephemeral-6983-2204/csi-hostpath-snapshotter-7ccf9ccb77\" objectUID=3a89c0a0-42f8-4860-9015-427f77d6ceb2 kind=\"ControllerRevision\" virtual=false\nI0718 20:00:37.295572       1 stateful_set.go:419] StatefulSet has been deleted ephemeral-6983-2204/csi-hostpath-snapshotter\nI0718 20:00:37.295614       1 garbagecollector.go:471] \"Processing object\" object=\"ephemeral-6983-2204/csi-hostpath-snapshotter-0\" objectUID=eaf14c55-6735-4e2a-802c-ca190733015b kind=\"Pod\" virtual=false\nI0718 20:00:37.302091       1 garbagecollector.go:580] \"Deleting object\" object=\"ephemeral-6983-2204/csi-hostpath-snapshotter-7ccf9ccb77\" objectUID=3a89c0a0-42f8-4860-9015-427f77d6ceb2 kind=\"ControllerRevision\" propagationPolicy=Background\nI0718 20:00:37.302207       1 garbagecollector.go:580] \"Deleting object\" object=\"ephemeral-6983-2204/csi-hostpath-snapshotter-0\" objectUID=eaf14c55-6735-4e2a-802c-ca190733015b kind=\"Pod\" propagationPolicy=Background\nI0718 20:00:37.504834       1 namespace_controller.go:185] Namespace has been deleted provisioning-6645\nI0718 20:00:37.575008       1 namespace_controller.go:185] Namespace has been deleted deployment-672\nI0718 20:00:37.710075       1 namespace_controller.go:185] Namespace has been deleted node-problem-detector-379\nI0718 20:00:38.845206       1 event.go:291] \"Event occurred\" object=\"csi-mock-volumes-5970/pvc-pw64q\" kind=\"PersistentVolumeClaim\" apiVersion=\"v1\" type=\"Normal\" reason=\"ExternalProvisioning\" message=\"waiting for a volume to be created, either by external provisioner \\\"csi-mock-csi-mock-volumes-5970\\\" or manually created by system administrator\"\nI0718 20:00:38.901932       1 pv_controller.go:879] volume \"pvc-471640f7-8ba5-4c6d-a059-e4a0fe35948f\" entered phase \"Bound\"\nI0718 20:00:38.901962       1 pv_controller.go:982] volume \"pvc-471640f7-8ba5-4c6d-a059-e4a0fe35948f\" bound to claim \"csi-mock-volumes-5970/pvc-pw64q\"\nI0718 20:00:38.912642       1 pv_controller.go:823] claim \"csi-mock-volumes-5970/pvc-pw64q\" entered phase \"Bound\"\nE0718 20:00:39.327504       1 reflector.go:138] k8s.io/client-go/metadata/metadatainformer/informer.go:90: Failed to watch *v1.PartialObjectMetadata: failed to list *v1.PartialObjectMetadata: the server could not find the requested resource\nI0718 20:00:39.482005       1 replica_set.go:559] \"Too few replicas\" replicaSet=\"services-8331/up-down-2\" need=3 creating=1\nI0718 20:00:39.517937       1 replica_set.go:559] \"Too few replicas\" replicaSet=\"services-8331/up-down-3\" need=3 creating=1\nI0718 20:00:39.556192       1 garbagecollector.go:471] \"Processing object\" object=\"services-8331/up-down-2-9llzs\" objectUID=420d9ca1-39df-4c4d-8204-0b64b7663485 kind=\"Pod\" virtual=false\nI0718 20:00:39.556213       1 garbagecollector.go:471] \"Processing object\" object=\"services-8331/up-down-2-7ll7x\" objectUID=7fd419a4-6782-4394-ac2e-e0eb61c32025 kind=\"Pod\" virtual=false\nI0718 20:00:39.556231       1 garbagecollector.go:471] \"Processing object\" object=\"services-8331/up-down-2-p6tjg\" objectUID=4f98aea4-64e6-456b-ad50-68f454d3443d kind=\"Pod\" virtual=false\nI0718 20:00:39.562417       1 garbagecollector.go:471] \"Processing object\" object=\"services-8331/up-down-3-mzjl7\" objectUID=56fbf12d-0af2-48c6-ba4f-901d5ebbcbf8 kind=\"Pod\" virtual=false\nI0718 20:00:39.562445       1 garbagecollector.go:471] \"Processing object\" object=\"services-8331/up-down-3-w5cct\" objectUID=4a8d31cb-803a-4a6d-abfb-84984f11f108 kind=\"Pod\" virtual=false\nI0718 20:00:39.562458       1 garbagecollector.go:471] \"Processing object\" object=\"services-8331/up-down-3-r7rk6\" objectUID=9e40e0ad-60fd-4c3f-8b7c-463bafcdb218 kind=\"Pod\" virtual=false\nI0718 20:00:39.572381       1 garbagecollector.go:471] \"Processing object\" object=\"services-8331/up-down-2-hzlnz\" objectUID=0fcea07a-947b-4a98-9e14-56b69ff9e4e5 kind=\"EndpointSlice\" virtual=false\nI0718 20:00:39.582241       1 garbagecollector.go:580] \"Deleting object\" object=\"services-8331/up-down-2-hzlnz\" objectUID=0fcea07a-947b-4a98-9e14-56b69ff9e4e5 kind=\"EndpointSlice\" propagationPolicy=Background\nI0718 20:00:39.588200       1 garbagecollector.go:471] \"Processing object\" object=\"services-8331/up-down-3-927k2\" objectUID=8af7cf05-0519-44e3-93a4-7aee5551f2ff kind=\"EndpointSlice\" virtual=false\nI0718 20:00:39.590982       1 garbagecollector.go:580] \"Deleting object\" object=\"services-8331/up-down-3-927k2\" objectUID=8af7cf05-0519-44e3-93a4-7aee5551f2ff kind=\"EndpointSlice\" propagationPolicy=Background\nI0718 20:00:40.286639       1 namespace_controller.go:185] Namespace has been deleted ephemeral-6983\nE0718 20:00:40.303131       1 namespace_controller.go:162] deletion of namespace services-8331 failed: unexpected items still remain in namespace: services-8331 for gvr: /v1, Resource=pods\nE0718 20:00:40.477693       1 namespace_controller.go:162] deletion of namespace services-8331 failed: unexpected items still remain in namespace: services-8331 for gvr: /v1, Resource=pods\nE0718 20:00:40.641580       1 namespace_controller.go:162] deletion of namespace services-8331 failed: unexpected items still remain in namespace: services-8331 for gvr: /v1, Resource=pods\nE0718 20:00:40.818167       1 namespace_controller.go:162] deletion of namespace services-8331 failed: unexpected items still remain in namespace: services-8331 for gvr: /v1, Resource=pods\nE0718 20:00:41.044707       1 namespace_controller.go:162] deletion of namespace services-8331 failed: unexpected items still remain in namespace: services-8331 for gvr: /v1, Resource=pods\nE0718 20:00:41.289709       1 namespace_controller.go:162] deletion of namespace services-8331 failed: unexpected items still remain in namespace: services-8331 for gvr: /v1, Resource=pods\nI0718 20:00:41.515994       1 namespace_controller.go:185] Namespace has been deleted provisioning-1326\nE0718 20:00:41.615737       1 namespace_controller.go:162] deletion of namespace services-8331 failed: unexpected items still remain in namespace: services-8331 for gvr: /v1, Resource=pods\nI0718 20:00:41.895589       1 pv_controller.go:879] volume \"local-pvp4h9c\" entered phase \"Available\"\nI0718 20:00:41.921459       1 pv_controller.go:930] claim \"persistent-local-volumes-test-2161/pvc-fdn29\" bound to volume \"local-pvp4h9c\"\nI0718 20:00:41.936247       1 pv_controller.go:879] volume \"local-pvp4h9c\" entered phase \"Bound\"\nI0718 20:00:41.936272       1 pv_controller.go:982] volume \"local-pvp4h9c\" bound to claim \"persistent-local-volumes-test-2161/pvc-fdn29\"\nI0718 20:00:41.947629       1 pv_controller.go:823] claim \"persistent-local-volumes-test-2161/pvc-fdn29\" entered phase \"Bound\"\nE0718 20:00:42.146648       1 namespace_controller.go:162] deletion of namespace services-8331 failed: unexpected items still remain in namespace: services-8331 for gvr: /v1, Resource=pods\nE0718 20:00:42.412962       1 tokens_controller.go:262] error synchronizing serviceaccount ephemeral-6983-2204/default: secrets \"default-token-jfn9q\" is forbidden: unable to create new content in namespace ephemeral-6983-2204 because it is being terminated\nE0718 20:00:42.542797       1 tokens_controller.go:262] error synchronizing serviceaccount volumemode-6763/default: secrets \"default-token-n74zq\" is forbidden: unable to create new content in namespace volumemode-6763 because it is being terminated\nI0718 20:00:42.823624       1 graph_builder.go:587] add [v1/Pod, namespace: ephemeral-7768, name: inline-volume-tester-jpmwb, uid: 3c44a6bf-44e0-4ac4-b590-7642709f626c] to the attemptToDelete, because it's waiting for its dependents to be deleted\nI0718 20:00:42.823696       1 garbagecollector.go:471] \"Processing object\" object=\"ephemeral-7768/inline-volume-tester-jpmwb\" objectUID=3c44a6bf-44e0-4ac4-b590-7642709f626c kind=\"Pod\" virtual=false\nI0718 20:00:42.827052       1 garbagecollector.go:590] remove DeleteDependents finalizer for item [v1/Pod, namespace: ephemeral-7768, name: inline-volume-tester-jpmwb, uid: 3c44a6bf-44e0-4ac4-b590-7642709f626c]\nE0718 20:00:42.952293       1 namespace_controller.go:162] deletion of namespace services-8331 failed: unexpected items still remain in namespace: services-8331 for gvr: /v1, Resource=pods\nE0718 20:00:43.387465       1 reflector.go:138] k8s.io/client-go/metadata/metadatainformer/informer.go:90: Failed to watch *v1.PartialObjectMetadata: failed to list *v1.PartialObjectMetadata: the server could not find the requested resource\nE0718 20:00:44.401645       1 namespace_controller.go:162] deletion of namespace services-8331 failed: unexpected items still remain in namespace: services-8331 for gvr: /v1, Resource=pods\nI0718 20:00:44.795309       1 namespace_controller.go:185] Namespace has been deleted provisioning-5909\nI0718 20:00:45.411317       1 pvc_protection_controller.go:291] \"PVC is unused\" PVC=\"provisioning-8250/pvc-ssjhx\"\nI0718 20:00:45.425770       1 pv_controller.go:640] volume \"local-fzxpt\" is released and reclaim policy \"Retain\" will be executed\nI0718 20:00:45.431907       1 pv_controller.go:879] volume \"local-fzxpt\" entered phase \"Released\"\nI0718 20:00:45.447965       1 pv_controller_base.go:505] deletion of claim \"provisioning-8250/pvc-ssjhx\" was already processed\nI0718 20:00:46.000784       1 deployment_controller.go:583] \"Deployment has been deleted\" deployment=\"webhook-8332/sample-webhook-deployment\"\nE0718 20:00:47.156127       1 namespace_controller.go:162] deletion of namespace services-8331 failed: unexpected items still remain in namespace: services-8331 for gvr: /v1, Resource=pods\nI0718 20:00:47.702274       1 event.go:291] \"Event occurred\" object=\"csi-mock-volumes-735-8114/csi-mockplugin\" kind=\"StatefulSet\" apiVersion=\"apps/v1\" type=\"Normal\" reason=\"SuccessfulCreate\" message=\"create Pod csi-mockplugin-0 in StatefulSet csi-mockplugin successful\"\nI0718 20:00:47.770748       1 event.go:291] \"Event occurred\" object=\"csi-mock-volumes-735-8114/csi-mockplugin-attacher\" kind=\"StatefulSet\" apiVersion=\"apps/v1\" type=\"Normal\" reason=\"SuccessfulCreate\" message=\"create Pod csi-mockplugin-attacher-0 in StatefulSet csi-mockplugin-attacher successful\"\nI0718 20:00:47.780196       1 namespace_controller.go:185] Namespace has been deleted volumemode-6763\nE0718 20:00:48.214413       1 reflector.go:138] k8s.io/client-go/metadata/metadatainformer/informer.go:90: Failed to watch *v1.PartialObjectMetadata: failed to list *v1.PartialObjectMetadata: the server could not find the requested resource\nI0718 20:00:50.669825       1 event.go:291] \"Event occurred\" object=\"ephemeral-5827/inline-volume-tester2-qw2vr-my-volume-0\" kind=\"PersistentVolumeClaim\" apiVersion=\"v1\" type=\"Normal\" reason=\"WaitForPodScheduled\" message=\"waiting for pod inline-volume-tester2-qw2vr to be scheduled\"\nE0718 20:00:51.429393       1 reflector.go:138] k8s.io/client-go/metadata/metadatainformer/informer.go:90: Failed to watch *v1.PartialObjectMetadata: failed to list *v1.PartialObjectMetadata: the server could not find the requested resource\nI0718 20:00:51.566867       1 event.go:291] \"Event occurred\" object=\"ephemeral-5827/inline-volume-tester2-qw2vr-my-volume-0\" kind=\"PersistentVolumeClaim\" apiVersion=\"v1\" type=\"Normal\" reason=\"WaitForPodScheduled\" message=\"waiting for pod inline-volume-tester2-qw2vr to be scheduled\"\nI0718 20:00:51.652194       1 pv_controller.go:879] volume \"local-pvwqfc2\" entered phase \"Available\"\nI0718 20:00:51.678469       1 pv_controller.go:930] claim \"persistent-local-volumes-test-7373/pvc-f9lvq\" bound to volume \"local-pvwqfc2\"\nI0718 20:00:51.696104       1 pv_controller.go:879] volume \"local-pvwqfc2\" entered phase \"Bound\"\nI0718 20:00:51.696125       1 pv_controller.go:982] volume \"local-pvwqfc2\" bound to claim \"persistent-local-volumes-test-7373/pvc-f9lvq\"\nI0718 20:00:51.708948       1 pv_controller.go:823] claim \"persistent-local-volumes-test-7373/pvc-f9lvq\" entered phase \"Bound\"\nI0718 20:00:51.781693       1 event.go:291] \"Event occurred\" object=\"ephemeral-5827/inline-volume-tester2-qw2vr-my-volume-0\" kind=\"PersistentVolumeClaim\" apiVersion=\"v1\" type=\"Normal\" reason=\"ExternalProvisioning\" message=\"waiting for a volume to be created, either by external provisioner \\\"csi-hostpath-ephemeral-5827\\\" or manually created by system administrator\"\nI0718 20:00:51.811248       1 pv_controller.go:879] volume \"pvc-cd083b17-8ca9-40cf-80e6-aefca09f6ef7\" entered phase \"Bound\"\nI0718 20:00:51.811280       1 pv_controller.go:982] volume \"pvc-cd083b17-8ca9-40cf-80e6-aefca09f6ef7\" bound to claim \"ephemeral-5827/inline-volume-tester2-qw2vr-my-volume-0\"\nI0718 20:00:51.822945       1 pv_controller.go:823] claim \"ephemeral-5827/inline-volume-tester2-qw2vr-my-volume-0\" entered phase \"Bound\"\nI0718 20:00:52.001623       1 namespace_controller.go:185] Namespace has been deleted ephemeral-2453-2953\nI0718 20:00:52.018165       1 event.go:291] \"Event occurred\" object=\"job-2919/backofflimit\" kind=\"Job\" apiVersion=\"batch/v1\" type=\"Warning\" reason=\"BackoffLimitExceeded\" message=\"Job has reached the specified backoff limit\"\nE0718 20:00:52.331320       1 namespace_controller.go:162] deletion of namespace services-8331 failed: unexpected items still remain in namespace: services-8331 for gvr: /v1, Resource=pods\nI0718 20:00:52.353820       1 replica_set.go:559] \"Too few replicas\" replicaSet=\"disruption-5763/rs\" need=10 creating=1\nI0718 20:00:52.360538       1 event.go:291] \"Event occurred\" object=\"disruption-5763/rs\" kind=\"ReplicaSet\" apiVersion=\"apps/v1\" type=\"Normal\" reason=\"SuccessfulCreate\" message=\"Created pod: rs-pg2g7\"\nI0718 20:00:52.789234       1 reconciler.go:295] attacherDetacher.AttachVolume started for volume \"pvc-cd083b17-8ca9-40cf-80e6-aefca09f6ef7\" (UniqueName: \"kubernetes.io/csi/csi-hostpath-ephemeral-5827^da7cbc4f-e802-11eb-8233-02fc2257e70c\") from node \"10.116.0.7\" \nI0718 20:00:53.069788       1 replica_set.go:559] \"Too few replicas\" replicaSet=\"apply-7798/deployment-585449566\" need=3 creating=3\nI0718 20:00:53.070122       1 event.go:291] \"Event occurred\" object=\"apply-7798/deployment\" kind=\"Deployment\" apiVersion=\"apps/v1\" type=\"Normal\" reason=\"ScalingReplicaSet\" message=\"Scaled up replica set deployment-585449566 to 3\"\nI0718 20:00:53.081462       1 event.go:291] \"Event occurred\" object=\"apply-7798/deployment-585449566\" kind=\"ReplicaSet\" apiVersion=\"apps/v1\" type=\"Normal\" reason=\"SuccessfulCreate\" message=\"Created pod: deployment-585449566-s7tzv\"\nI0718 20:00:53.082002       1 deployment_controller.go:490] \"Error syncing deployment\" deployment=\"apply-7798/deployment\" err=\"Operation cannot be fulfilled on deployments.apps \\\"deployment\\\": the object has been modified; please apply your changes to the latest version and try again\"\nI0718 20:00:53.088965       1 event.go:291] \"Event occurred\" object=\"apply-7798/deployment-585449566\" kind=\"ReplicaSet\" apiVersion=\"apps/v1\" type=\"Normal\" reason=\"SuccessfulCreate\" message=\"Created pod: deployment-585449566-kfjwp\"\nI0718 20:00:53.089083       1 event.go:291] \"Event occurred\" object=\"apply-7798/deployment-585449566\" kind=\"ReplicaSet\" apiVersion=\"apps/v1\" type=\"Normal\" reason=\"SuccessfulCreate\" message=\"Created pod: deployment-585449566-v9pgj\"\nI0718 20:00:53.113135       1 replica_set.go:559] \"Too few replicas\" replicaSet=\"apply-7798/deployment-55649fd747\" need=1 creating=1\nI0718 20:00:53.113161       1 event.go:291] \"Event occurred\" object=\"apply-7798/deployment\" kind=\"Deployment\" apiVersion=\"apps/v1\" type=\"Normal\" reason=\"ScalingReplicaSet\" message=\"Scaled up replica set deployment-55649fd747 to 1\"\nI0718 20:00:53.118925       1 event.go:291] \"Event occurred\" object=\"apply-7798/deployment-55649fd747\" kind=\"ReplicaSet\" apiVersion=\"apps/v1\" type=\"Normal\" reason=\"SuccessfulCreate\" message=\"Created pod: deployment-55649fd747-n2j8d\"\nI0718 20:00:53.132868       1 deployment_controller.go:490] \"Error syncing deployment\" deployment=\"apply-7798/deployment\" err=\"Operation cannot be fulfilled on deployments.apps \\\"deployment\\\": the object has been modified; please apply your changes to the latest version and try again\"\nI0718 20:00:53.244496       1 garbagecollector.go:471] \"Processing object\" object=\"apply-7798/deployment-585449566\" objectUID=f6223690-3f07-4a64-a123-50530d9581f0 kind=\"ReplicaSet\" virtual=false\nI0718 20:00:53.244513       1 garbagecollector.go:471] \"Processing object\" object=\"apply-7798/deployment-55649fd747\" objectUID=2e08d9b4-59c9-42f8-a5e6-297dd4e89788 kind=\"ReplicaSet\" virtual=false\nI0718 20:00:53.244518       1 deployment_controller.go:583] \"Deployment has been deleted\" deployment=\"apply-7798/deployment\"\nI0718 20:00:53.247363       1 garbagecollector.go:580] \"Deleting object\" object=\"apply-7798/deployment-55649fd747\" objectUID=2e08d9b4-59c9-42f8-a5e6-297dd4e89788 kind=\"ReplicaSet\" propagationPolicy=Background\nI0718 20:00:53.248273       1 garbagecollector.go:580] \"Deleting object\" object=\"apply-7798/deployment-585449566\" objectUID=f6223690-3f07-4a64-a123-50530d9581f0 kind=\"ReplicaSet\" propagationPolicy=Background\nI0718 20:00:53.253314       1 garbagecollector.go:471] \"Processing object\" object=\"apply-7798/deployment-55649fd747-n2j8d\" objectUID=3795dbfc-307a-42e1-9661-a37e6cc86ffe kind=\"Pod\" virtual=false\nI0718 20:00:53.256331       1 garbagecollector.go:471] \"Processing object\" object=\"apply-7798/deployment-585449566-kfjwp\" objectUID=bb082671-c11f-4299-b160-8532c9d9ead7 kind=\"Pod\" virtual=false\nI0718 20:00:53.256377       1 garbagecollector.go:471] \"Processing object\" object=\"apply-7798/deployment-585449566-v9pgj\" objectUID=afd2492f-33c3-4976-8292-7f94e8821383 kind=\"Pod\" virtual=false\nI0718 20:00:53.256334       1 garbagecollector.go:471] \"Processing object\" object=\"apply-7798/deployment-585449566-s7tzv\" objectUID=9d2e2d85-732b-4cd9-9aab-a2ddb390b313 kind=\"Pod\" virtual=false\nI0718 20:00:53.256488       1 garbagecollector.go:580] \"Deleting object\" object=\"apply-7798/deployment-55649fd747-n2j8d\" objectUID=3795dbfc-307a-42e1-9661-a37e6cc86ffe kind=\"Pod\" propagationPolicy=Background\nI0718 20:00:53.259196       1 garbagecollector.go:580] \"Deleting object\" object=\"apply-7798/deployment-585449566-kfjwp\" objectUID=bb082671-c11f-4299-b160-8532c9d9ead7 kind=\"Pod\" propagationPolicy=Background\nI0718 20:00:53.260090       1 garbagecollector.go:580] \"Deleting object\" object=\"apply-7798/deployment-585449566-s7tzv\" objectUID=9d2e2d85-732b-4cd9-9aab-a2ddb390b313 kind=\"Pod\" propagationPolicy=Background\nI0718 20:00:53.260346       1 garbagecollector.go:580] \"Deleting object\" object=\"apply-7798/deployment-585449566-v9pgj\" objectUID=afd2492f-33c3-4976-8292-7f94e8821383 kind=\"Pod\" propagationPolicy=Background\nI0718 20:00:53.373353       1 operation_generator.go:368] AttachVolume.Attach succeeded for volume \"pvc-cd083b17-8ca9-40cf-80e6-aefca09f6ef7\" (UniqueName: \"kubernetes.io/csi/csi-hostpath-ephemeral-5827^da7cbc4f-e802-11eb-8233-02fc2257e70c\") from node \"10.116.0.7\" \nI0718 20:00:53.373504       1 event.go:291] \"Event occurred\" object=\"ephemeral-5827/inline-volume-tester2-qw2vr\" kind=\"Pod\" apiVersion=\"v1\" type=\"Normal\" reason=\"SuccessfulAttachVolume\" message=\"AttachVolume.Attach succeeded for volume \\\"pvc-cd083b17-8ca9-40cf-80e6-aefca09f6ef7\\\" \"\nI0718 20:00:53.873818       1 pvc_protection_controller.go:291] \"PVC is unused\" PVC=\"persistent-local-volumes-test-7373/pvc-f9lvq\"\nI0718 20:00:53.885142       1 pv_controller.go:640] volume \"local-pvwqfc2\" is released and reclaim policy \"Retain\" will be executed\nI0718 20:00:53.889644       1 pv_controller.go:879] volume \"local-pvwqfc2\" entered phase \"Released\"\nI0718 20:00:53.914913       1 pv_controller_base.go:505] deletion of claim \"persistent-local-volumes-test-7373/pvc-f9lvq\" was already processed\nE0718 20:00:54.465641       1 tokens_controller.go:262] error synchronizing serviceaccount ephemeral-8120/default: secrets \"default-token-5z26x\" is forbidden: unable to create new content in namespace ephemeral-8120 because it is being terminated\nE0718 20:00:54.830833       1 tokens_controller.go:262] error synchronizing serviceaccount svcaccounts-575/default: secrets \"default-token-85h46\" is forbidden: unable to create new content in namespace svcaccounts-575 because it is being terminated\nI0718 20:00:54.947315       1 garbagecollector.go:471] \"Processing object\" object=\"cronjob-8504/forbid-27110640\" objectUID=62c1373a-ab51-4d26-acd7-40b5af2e95e3 kind=\"Job\" virtual=false\nI0718 20:00:54.949722       1 garbagecollector.go:580] \"Deleting object\" object=\"cronjob-8504/forbid-27110640\" objectUID=62c1373a-ab51-4d26-acd7-40b5af2e95e3 kind=\"Job\" propagationPolicy=Background\nI0718 20:00:54.960325       1 garbagecollector.go:471] \"Processing object\" object=\"cronjob-8504/forbid-27110640-q9dv4\" objectUID=46778522-479e-4b8c-97dc-a27301147d2d kind=\"Pod\" virtual=false\nI0718 20:00:54.965417       1 garbagecollector.go:580] \"Deleting object\" object=\"cronjob-8504/forbid-27110640-q9dv4\" objectUID=46778522-479e-4b8c-97dc-a27301147d2d kind=\"Pod\" propagationPolicy=Background\nI0718 20:00:55.728484       1 event.go:291] \"Event occurred\" object=\"provisioning-1806-9608/csi-hostpath-attacher\" kind=\"StatefulSet\" apiVersion=\"apps/v1\" type=\"Normal\" reason=\"SuccessfulCreate\" message=\"create Pod csi-hostpath-attacher-0 in StatefulSet csi-hostpath-attacher successful\"\nI0718 20:00:55.809844       1 event.go:291] \"Event occurred\" object=\"provisioning-1806-9608/csi-hostpathplugin\" kind=\"StatefulSet\" apiVersion=\"apps/v1\" type=\"Normal\" reason=\"SuccessfulCreate\" message=\"create Pod csi-hostpathplugin-0 in StatefulSet csi-hostpathplugin successful\"\nI0718 20:00:55.895508       1 event.go:291] \"Event occurred\" object=\"provisioning-1806-9608/csi-hostpath-provisioner\" kind=\"StatefulSet\" apiVersion=\"apps/v1\" type=\"Normal\" reason=\"SuccessfulCreate\" message=\"create Pod csi-hostpath-provisioner-0 in StatefulSet csi-hostpath-provisioner successful\"\nI0718 20:00:55.964468       1 event.go:291] \"Event occurred\" object=\"provisioning-1806-9608/csi-hostpath-resizer\" kind=\"StatefulSet\" apiVersion=\"apps/v1\" type=\"Normal\" reason=\"SuccessfulCreate\" message=\"create Pod csi-hostpath-resizer-0 in StatefulSet csi-hostpath-resizer successful\"\nI0718 20:00:56.038446       1 event.go:291] \"Event occurred\" object=\"provisioning-1806-9608/csi-hostpath-snapshotter\" kind=\"StatefulSet\" apiVersion=\"apps/v1\" type=\"Normal\" reason=\"SuccessfulCreate\" message=\"create Pod csi-hostpath-snapshotter-0 in StatefulSet csi-hostpath-snapshotter successful\"\nI0718 20:00:56.132250       1 event.go:291] \"Event occurred\" object=\"provisioning-1806/csi-hostpath4ndks\" kind=\"PersistentVolumeClaim\" apiVersion=\"v1\" type=\"Normal\" reason=\"ExternalProvisioning\" message=\"waiting for a volume to be created, either by external provisioner \\\"csi-hostpath-provisioning-1806\\\" or manually created by system administrator\"\nI0718 20:00:56.132271       1 event.go:291] \"Event occurred\" object=\"provisioning-1806/csi-hostpath4ndks\" kind=\"PersistentVolumeClaim\" apiVersion=\"v1\" type=\"Normal\" reason=\"ExternalProvisioning\" message=\"waiting for a volume to be created, either by external provisioner \\\"csi-hostpath-provisioning-1806\\\" or manually created by system administrator\"\nI0718 20:00:56.190506       1 namespace_controller.go:185] Namespace has been deleted provisioning-8250\nI0718 20:00:56.461526       1 garbagecollector.go:471] \"Processing object\" object=\"ephemeral-8120-5728/csi-hostpath-attacher-gvhr6\" objectUID=8d7d9576-95df-4034-9f5f-85a9817aa529 kind=\"EndpointSlice\" virtual=false\nI0718 20:00:56.469227       1 garbagecollector.go:580] \"Deleting object\" object=\"ephemeral-8120-5728/csi-hostpath-attacher-gvhr6\" objectUID=8d7d9576-95df-4034-9f5f-85a9817aa529 kind=\"EndpointSlice\" propagationPolicy=Background\nI0718 20:00:56.515743       1 stateful_set.go:419] StatefulSet has been deleted ephemeral-8120-5728/csi-hostpath-attacher\nI0718 20:00:56.515757       1 garbagecollector.go:471] \"Processing object\" object=\"ephemeral-8120-5728/csi-hostpath-attacher-0\" objectUID=edf3466e-b798-417f-97c3-a8b2eb2b857f kind=\"Pod\" virtual=false\nI0718 20:00:56.515757       1 garbagecollector.go:471] \"Processing object\" object=\"ephemeral-8120-5728/csi-hostpath-attacher-7d9599c87\" objectUID=72acd0c5-1bed-4609-bfe1-b4267e560cdb kind=\"ControllerRevision\" virtual=false\nI0718 20:00:56.518242       1 garbagecollector.go:580] \"Deleting object\" object=\"ephemeral-8120-5728/csi-hostpath-attacher-7d9599c87\" objectUID=72acd0c5-1bed-4609-bfe1-b4267e560cdb kind=\"ControllerRevision\" propagationPolicy=Background\nI0718 20:00:56.519026       1 garbagecollector.go:580] \"Deleting object\" object=\"ephemeral-8120-5728/csi-hostpath-attacher-0\" objectUID=edf3466e-b798-417f-97c3-a8b2eb2b857f kind=\"Pod\" propagationPolicy=Background\nI0718 20:00:56.837497       1 garbagecollector.go:471] \"Processing object\" object=\"ephemeral-8120-5728/csi-hostpathplugin-mrr5q\" objectUID=ae8ec62e-2452-467a-b249-4e08cc47ed59 kind=\"EndpointSlice\" virtual=false\nI0718 20:00:56.842791       1 garbagecollector.go:580] \"Deleting object\" object=\"ephemeral-8120-5728/csi-hostpathplugin-mrr5q\" objectUID=ae8ec62e-2452-467a-b249-4e08cc47ed59 kind=\"EndpointSlice\" propagationPolicy=Background\nI0718 20:00:56.885348       1 garbagecollector.go:471] \"Processing object\" object=\"ephemeral-8120-5728/csi-hostpathplugin-7b66d895c7\" objectUID=5f08419e-fa03-4099-a6e9-2e8446777784 kind=\"ControllerRevision\" virtual=false\nI0718 20:00:56.885402       1 stateful_set.go:419] StatefulSet has been deleted ephemeral-8120-5728/csi-hostpathplugin\nI0718 20:00:56.885444       1 garbagecollector.go:471] \"Processing object\" object=\"ephemeral-8120-5728/csi-hostpathplugin-0\" objectUID=3b9c3ad9-dbe1-4217-93c7-d3f24da432ac kind=\"Pod\" virtual=false\nI0718 20:00:56.887683       1 garbagecollector.go:580] \"Deleting object\" object=\"ephemeral-8120-5728/csi-hostpathplugin-7b66d895c7\" objectUID=5f08419e-fa03-4099-a6e9-2e8446777784 kind=\"ControllerRevision\" propagationPolicy=Background\nI0718 20:00:56.888589       1 garbagecollector.go:580] \"Deleting object\" object=\"ephemeral-8120-5728/csi-hostpathplugin-0\" objectUID=3b9c3ad9-dbe1-4217-93c7-d3f24da432ac kind=\"Pod\" propagationPolicy=Background\nI0718 20:00:56.918691       1 garbagecollector.go:471] \"Processing object\" object=\"ephemeral-8120-5728/csi-hostpath-provisioner-g58ff\" objectUID=5f80a9c8-29b1-4808-a36c-dae001a7fd49 kind=\"EndpointSlice\" virtual=false\nI0718 20:00:56.930077       1 garbagecollector.go:580] \"Deleting object\" object=\"ephemeral-8120-5728/csi-hostpath-provisioner-g58ff\" objectUID=5f80a9c8-29b1-4808-a36c-dae001a7fd49 kind=\"EndpointSlice\" propagationPolicy=Background\nI0718 20:00:56.976851       1 garbagecollector.go:471] \"Processing object\" object=\"ephemeral-8120-5728/csi-hostpath-provisioner-668688785\" objectUID=b58490c4-4b78-4719-99e5-5af8031f601e kind=\"ControllerRevision\" virtual=false\nI0718 20:00:56.976885       1 stateful_set.go:419] StatefulSet has been deleted ephemeral-8120-5728/csi-hostpath-provisioner\nI0718 20:00:56.976941       1 garbagecollector.go:471] \"Processing object\" object=\"ephemeral-8120-5728/csi-hostpath-provisioner-0\" objectUID=fdc247c6-598a-49d1-a838-6eebc9f60857 kind=\"Pod\" virtual=false\nI0718 20:00:56.980113       1 garbagecollector.go:580] \"Deleting object\" object=\"ephemeral-8120-5728/csi-hostpath-provisioner-668688785\" objectUID=b58490c4-4b78-4719-99e5-5af8031f601e kind=\"ControllerRevision\" propagationPolicy=Background\nI0718 20:00:56.980892       1 garbagecollector.go:580] \"Deleting object\" object=\"ephemeral-8120-5728/csi-hostpath-provisioner-0\" objectUID=fdc247c6-598a-49d1-a838-6eebc9f60857 kind=\"Pod\" propagationPolicy=Background\nI0718 20:00:57.010269       1 garbagecollector.go:471] \"Processing object\" object=\"ephemeral-8120-5728/csi-hostpath-resizer-snd78\" objectUID=69d46928-0d82-4bd5-8d27-8c7acae4a275 kind=\"EndpointSlice\" virtual=false\nI0718 20:00:57.013018       1 garbagecollector.go:580] \"Deleting object\" object=\"ephemeral-8120-5728/csi-hostpath-resizer-snd78\" objectUID=69d46928-0d82-4bd5-8d27-8c7acae4a275 kind=\"EndpointSlice\" propagationPolicy=Background\nI0718 20:00:57.065188       1 garbagecollector.go:471] \"Processing object\" object=\"ephemeral-8120-5728/csi-hostpath-resizer-7445995556\" objectUID=ac53b20d-9974-4a6e-a7d5-01b84ab024e8 kind=\"ControllerRevision\" virtual=false\nI0718 20:00:57.065266       1 stateful_set.go:419] StatefulSet has been deleted ephemeral-8120-5728/csi-hostpath-resizer\nI0718 20:00:57.065323       1 garbagecollector.go:471] \"Processing object\" object=\"ephemeral-8120-5728/csi-hostpath-resizer-0\" objectUID=c96889b1-caeb-47a4-9de0-6fc72a49bec0 kind=\"Pod\" virtual=false\nI0718 20:00:57.067827       1 garbagecollector.go:580] \"Deleting object\" object=\"ephemeral-8120-5728/csi-hostpath-resizer-0\" objectUID=c96889b1-caeb-47a4-9de0-6fc72a49bec0 kind=\"Pod\" propagationPolicy=Background\nI0718 20:00:57.068525       1 garbagecollector.go:580] \"Deleting object\" object=\"ephemeral-8120-5728/csi-hostpath-resizer-7445995556\" objectUID=ac53b20d-9974-4a6e-a7d5-01b84ab024e8 kind=\"ControllerRevision\" propagationPolicy=Background\nI0718 20:00:57.099439       1 garbagecollector.go:471] \"Processing object\" object=\"ephemeral-8120-5728/csi-hostpath-snapshotter-jwm2w\" objectUID=f04b1044-54ce-4b13-92d1-bfb3b1132977 kind=\"EndpointSlice\" virtual=false\nI0718 20:00:57.105194       1 garbagecollector.go:580] \"Deleting object\" object=\"ephemeral-8120-5728/csi-hostpath-snapshotter-jwm2w\" objectUID=f04b1044-54ce-4b13-92d1-bfb3b1132977 kind=\"EndpointSlice\" propagationPolicy=Background\nI0718 20:00:57.152716       1 garbagecollector.go:471] \"Processing object\" object=\"ephemeral-8120-5728/csi-hostpath-snapshotter-6b95b55868\" objectUID=cd9f9031-8649-4d78-8bf3-d6b707c444aa kind=\"ControllerRevision\" virtual=false\nI0718 20:00:57.152830       1 stateful_set.go:419] StatefulSet has been deleted ephemeral-8120-5728/csi-hostpath-snapshotter\nI0718 20:00:57.152864       1 garbagecollector.go:471] \"Processing object\" object=\"ephemeral-8120-5728/csi-hostpath-snapshotter-0\" objectUID=a2d6d832-a1e0-4f4f-96d1-d4eb74363c28 kind=\"Pod\" virtual=false\nI0718 20:00:57.155002       1 garbagecollector.go:580] \"Deleting object\" object=\"ephemeral-8120-5728/csi-hostpath-snapshotter-6b95b55868\" objectUID=cd9f9031-8649-4d78-8bf3-d6b707c444aa kind=\"ControllerRevision\" propagationPolicy=Background\nI0718 20:00:57.155847       1 garbagecollector.go:580] \"Deleting object\" object=\"ephemeral-8120-5728/csi-hostpath-snapshotter-0\" objectUID=a2d6d832-a1e0-4f4f-96d1-d4eb74363c28 kind=\"Pod\" propagationPolicy=Background\nI0718 20:00:57.472812       1 replica_set.go:559] \"Too few replicas\" replicaSet=\"disruption-5763/rs\" need=10 creating=1\nE0718 20:00:57.611747       1 namespace_controller.go:162] deletion of namespace services-8331 failed: unexpected items still remain in namespace: services-8331 for gvr: /v1, Resource=pods\nI0718 20:00:57.655594       1 garbagecollector.go:471] \"Processing object\" object=\"disruption-5763/rs-7rvh6\" objectUID=95a094f7-2280-4781-a060-c088aa8564dc kind=\"Pod\" virtual=false\nI0718 20:00:57.655609       1 garbagecollector.go:471] \"Processing object\" object=\"disruption-5763/rs-sgqwg\" objectUID=e833b424-99bd-4ecd-b0a2-92782d248114 kind=\"Pod\" virtual=false\nI0718 20:00:57.655622       1 garbagecollector.go:471] \"Processing object\" object=\"disruption-5763/rs-7pflq\" objectUID=a50e4ce3-5d39-48e1-a583-b96083aac0e0 kind=\"Pod\" virtual=false\nI0718 20:00:57.655627       1 garbagecollector.go:471] \"Processing object\" object=\"disruption-5763/rs-hbskc\" objectUID=f817e935-5e47-429b-9d5e-f6e0a79db0d9 kind=\"Pod\" virtual=false\nI0718 20:00:57.655633       1 garbagecollector.go:471] \"Processing object\" object=\"disruption-5763/rs-w2cw5\" objectUID=1454e7f7-f253-444e-aee9-5a6a9b4098a4 kind=\"Pod\" virtual=false\nI0718 20:00:57.655638       1 garbagecollector.go:471] \"Processing object\" object=\"disruption-5763/rs-ppfgl\" objectUID=921c444b-3699-4c86-9151-61a1357f6fbc kind=\"Pod\" virtual=false\nI0718 20:00:57.655650       1 garbagecollector.go:471] \"Processing object\" object=\"disruption-5763/rs-qm2bg\" objectUID=4dd810c2-b0bc-422e-b492-b13897a18f87 kind=\"Pod\" virtual=false\nI0718 20:00:57.655659       1 garbagecollector.go:471] \"Processing object\" object=\"disruption-5763/rs-2hk69\" objectUID=0c9c0ab4-f346-4210-8f21-be3ab74caa67 kind=\"Pod\" virtual=false\nI0718 20:00:57.655668       1 garbagecollector.go:471] \"Processing object\" object=\"disruption-5763/rs-pg2g7\" objectUID=8f244c37-4c84-468d-99e0-6872735787df kind=\"Pod\" virtual=false\nI0718 20:00:57.655669       1 garbagecollector.go:471] \"Processing object\" object=\"disruption-5763/rs-62n2z\" objectUID=8b386432-f002-49f3-b795-3a7ec18d59ad kind=\"Pod\" virtual=false\nI0718 20:00:57.655678       1 garbagecollector.go:471] \"Processing object\" object=\"disruption-5763/rs-67rmb\" objectUID=d16c7f62-c66e-4c0e-a8cb-266b477e3506 kind=\"Pod\" virtual=false\nI0718 20:00:57.990765       1 event.go:291] \"Event occurred\" object=\"csi-mock-volumes-735/pvc-nwzkz\" kind=\"PersistentVolumeClaim\" apiVersion=\"v1\" type=\"Normal\" reason=\"WaitForFirstConsumer\" message=\"waiting for first consumer to be created before binding\"\nI0718 20:00:58.039529       1 event.go:291] \"Event occurred\" object=\"csi-mock-volumes-735/pvc-nwzkz\" kind=\"PersistentVolumeClaim\" apiVersion=\"v1\" type=\"Normal\" reason=\"ExternalProvisioning\" message=\"waiting for a volume to be created, either by external provisioner \\\"csi-mock-csi-mock-volumes-735\\\" or manually created by system administrator\"\nI0718 20:00:58.059980       1 pv_controller.go:879] volume \"pvc-16ff8267-3caf-4d13-92fb-0d8c5cf3bed0\" entered phase \"Bound\"\nI0718 20:00:58.060015       1 pv_controller.go:982] volume \"pvc-16ff8267-3caf-4d13-92fb-0d8c5cf3bed0\" bound to claim \"csi-mock-volumes-735/pvc-nwzkz\"\nI0718 20:00:58.070813       1 pv_controller.go:823] claim \"csi-mock-volumes-735/pvc-nwzkz\" entered phase \"Bound\"\nI0718 20:00:58.195166       1 pv_controller.go:879] volume \"pvc-919825ea-7889-48f9-9d20-64716ca37948\" entered phase \"Bound\"\nI0718 20:00:58.195206       1 pv_controller.go:982] volume \"pvc-919825ea-7889-48f9-9d20-64716ca37948\" bound to claim \"provisioning-1806/csi-hostpath4ndks\"\nI0718 20:00:58.205202       1 pv_controller.go:823] claim \"provisioning-1806/csi-hostpath4ndks\" entered phase \"Bound\"\nE0718 20:00:58.249173       1 tokens_controller.go:262] error synchronizing serviceaccount apply-7798/default: secrets \"default-token-9mr4s\" is forbidden: unable to create new content in namespace apply-7798 because it is being terminated\nI0718 20:00:58.548826       1 namespace_controller.go:185] Namespace has been deleted init-container-65\nI0718 20:00:59.043035       1 reconciler.go:295] attacherDetacher.AttachVolume started for volume \"pvc-16ff8267-3caf-4d13-92fb-0d8c5cf3bed0\" (UniqueName: \"kubernetes.io/csi/csi-mock-csi-mock-volumes-735^4\") from node \"10.116.0.7\" \nE0718 20:00:59.371145       1 tokens_controller.go:262] error synchronizing serviceaccount persistent-local-volumes-test-7373/default: secrets \"default-token-tqfdn\" is forbidden: unable to create new content in namespace persistent-local-volumes-test-7373 because it is being terminated\nI0718 20:00:59.409895       1 namespace_controller.go:185] Namespace has been deleted pods-1715\nE0718 20:00:59.452947       1 reflector.go:138] k8s.io/client-go/metadata/metadatainformer/informer.go:90: Failed to watch *v1.PartialObjectMetadata: failed to list *v1.PartialObjectMetadata: the server could not find the requested resource\nI0718 20:00:59.571262       1 namespace_controller.go:185] Namespace has been deleted ephemeral-8120\nI0718 20:00:59.612600       1 operation_generator.go:368] AttachVolume.Attach succeeded for volume \"pvc-16ff8267-3caf-4d13-92fb-0d8c5cf3bed0\" (UniqueName: \"kubernetes.io/csi/csi-mock-csi-mock-volumes-735^4\") from node \"10.116.0.7\" \nI0718 20:00:59.612772       1 event.go:291] \"Event occurred\" object=\"csi-mock-volumes-735/pvc-volume-tester-h4m8q\" kind=\"Pod\" apiVersion=\"v1\" type=\"Normal\" reason=\"SuccessfulAttachVolume\" message=\"AttachVolume.Attach succeeded for volume \\\"pvc-16ff8267-3caf-4d13-92fb-0d8c5cf3bed0\\\" \"\nI0718 20:01:00.110130       1 namespace_controller.go:185] Namespace has been deleted svcaccounts-575\nE0718 20:01:00.115542       1 tokens_controller.go:262] error synchronizing serviceaccount cronjob-8504/default: secrets \"default-token-5crmg\" is forbidden: unable to create new content in namespace cronjob-8504 because it is being terminated\nI0718 20:01:00.119303       1 event.go:291] \"Event occurred\" object=\"cronjob-4272/concurrent\" kind=\"CronJob\" apiVersion=\"batch/v1\" type=\"Normal\" reason=\"SuccessfulCreate\" message=\"Created job concurrent-27110641\"\nI0718 20:01:00.132544       1 event.go:291] \"Event occurred\" object=\"cronjob-4272/concurrent-27110641\" kind=\"Job\" apiVersion=\"batch/v1\" type=\"Normal\" reason=\"SuccessfulCreate\" message=\"Created pod: concurrent-27110641-zx4v2\"\nI0718 20:01:00.132545       1 cronjob_controllerv2.go:193] \"error cleaning up jobs\" cronjob=\"cronjob-4272/concurrent\" resourceVersion=\"29800\" err=\"Operation cannot be fulfilled on cronjobs.batch \\\"concurrent\\\": the object has been modified; please apply your changes to the latest version and try again\"\nE0718 20:01:00.132570       1 cronjob_controllerv2.go:154] error syncing CronJobController cronjob-4272/concurrent, requeuing: Operation cannot be fulfilled on cronjobs.batch \"concurrent\": the object has been modified; please apply your changes to the latest version and try again\nI0718 20:01:00.361633       1 reconciler.go:295] attacherDetacher.AttachVolume started for volume \"pvc-919825ea-7889-48f9-9d20-64716ca37948\" (UniqueName: \"kubernetes.io/csi/csi-hostpath-provisioning-1806^de487053-e802-11eb-afb0-3e205f8df9a8\") from node \"10.116.0.5\" \nI0718 20:01:00.897368       1 event.go:291] \"Event occurred\" object=\"provisioning-1806/pod-subpath-test-dynamicpv-xb7h\" kind=\"Pod\" apiVersion=\"v1\" type=\"Normal\" reason=\"SuccessfulAttachVolume\" message=\"AttachVolume.Attach succeeded for volume \\\"pvc-919825ea-7889-48f9-9d20-64716ca37948\\\" \"\nI0718 20:01:00.897367       1 operation_generator.go:368] AttachVolume.Attach succeeded for volume \"pvc-919825ea-7889-48f9-9d20-64716ca37948\" (UniqueName: \"kubernetes.io/csi/csi-hostpath-provisioning-1806^de487053-e802-11eb-afb0-3e205f8df9a8\") from node \"10.116.0.5\" \nI0718 20:01:01.591780       1 reconciler.go:219] attacherDetacher.DetachVolume started for volume \"pvc-1522f266-6376-4263-9b6a-b0d2f01cbf7d\" (UniqueName: \"kubernetes.io/csi/csi-hostpath-provisioning-4462^ba5ecceb-e802-11eb-b14a-a6ffd9dc2bdc\") on node \"10.116.0.5\" \nI0718 20:01:01.594757       1 operation_generator.go:1483] Verified volume is safe to detach for volume \"pvc-1522f266-6376-4263-9b6a-b0d2f01cbf7d\" (UniqueName: \"kubernetes.io/csi/csi-hostpath-provisioning-4462^ba5ecceb-e802-11eb-b14a-a6ffd9dc2bdc\") on node \"10.116.0.5\" \nE0718 20:01:02.076228       1 tokens_controller.go:262] error synchronizing serviceaccount disruption-2-1142/default: secrets \"default-token-vmqk7\" is forbidden: unable to create new content in namespace disruption-2-1142 because it is being terminated\nE0718 20:01:02.155435       1 tokens_controller.go:262] error synchronizing serviceaccount disruption-667/default: secrets \"default-token-9cbzr\" is forbidden: unable to create new content in namespace disruption-667 because it is being terminated\nI0718 20:01:02.181928       1 operation_generator.go:483] DetachVolume.Detach succeeded for volume \"pvc-1522f266-6376-4263-9b6a-b0d2f01cbf7d\" (UniqueName: \"kubernetes.io/csi/csi-hostpath-provisioning-4462^ba5ecceb-e802-11eb-b14a-a6ffd9dc2bdc\") on node \"10.116.0.5\" \nI0718 20:01:02.199211       1 event.go:291] \"Event occurred\" object=\"provisioning-4462/pvc-bdxmk\" kind=\"PersistentVolumeClaim\" apiVersion=\"v1\" type=\"Normal\" reason=\"ExternalProvisioning\" message=\"waiting for a volume to be created, either by external provisioner \\\"csi-hostpath-provisioning-4462\\\" or manually created by system administrator\"\nI0718 20:01:02.199895       1 event.go:291] \"Event occurred\" object=\"provisioning-4462/pvc-bdxmk\" kind=\"PersistentVolumeClaim\" apiVersion=\"v1\" type=\"Normal\" reason=\"ExternalProvisioning\" message=\"waiting for a volume to be created, either by external provisioner \\\"csi-hostpath-provisioning-4462\\\" or manually created by system administrator\"\nI0718 20:01:02.243905       1 pv_controller.go:879] volume \"pvc-78fc2fe1-4514-4aa4-9fcf-c451cccbfb5a\" entered phase \"Bound\"\nI0718 20:01:02.243940       1 pv_controller.go:982] volume \"pvc-78fc2fe1-4514-4aa4-9fcf-c451cccbfb5a\" bound to claim \"provisioning-4462/pvc-bdxmk\"\nI0718 20:01:02.258304       1 pv_controller.go:823] claim \"provisioning-4462/pvc-bdxmk\" entered phase \"Bound\"\nE0718 20:01:02.642275       1 reflector.go:138] k8s.io/client-go/metadata/metadatainformer/informer.go:90: Failed to watch *v1.PartialObjectMetadata: failed to list *v1.PartialObjectMetadata: the server could not find the requested resource\nE0718 20:01:02.773114       1 namespace_controller.go:162] deletion of namespace services-8331 failed: unexpected items still remain in namespace: services-8331 for gvr: /v1, Resource=pods\nI0718 20:01:03.167325       1 namespace_controller.go:185] Namespace has been deleted job-2919\nI0718 20:01:03.806534       1 reconciler.go:295] attacherDetacher.AttachVolume started for volume \"pvc-78fc2fe1-4514-4aa4-9fcf-c451cccbfb5a\" (UniqueName: \"kubernetes.io/csi/csi-hostpath-provisioning-4462^e0b2c93a-e802-11eb-b14a-a6ffd9dc2bdc\") from node \"10.116.0.5\" \nI0718 20:01:03.918762       1 pv_controller.go:879] volume \"nfs-68cf7\" entered phase \"Available\"\nI0718 20:01:03.953263       1 pv_controller.go:930] claim \"pv-3628/pvc-wkzch\" bound to volume \"nfs-68cf7\"\nI0718 20:01:03.963762       1 pv_controller.go:879] volume \"nfs-68cf7\" entered phase \"Bound\"\nI0718 20:01:03.963787       1 pv_controller.go:982] volume \"nfs-68cf7\" bound to claim \"pv-3628/pvc-wkzch\"\nI0718 20:01:03.979250       1 pv_controller.go:823] claim \"pv-3628/pvc-wkzch\" entered phase \"Bound\"\nI0718 20:01:04.373736       1 operation_generator.go:368] AttachVolume.Attach succeeded for volume \"pvc-78fc2fe1-4514-4aa4-9fcf-c451cccbfb5a\" (UniqueName: \"kubernetes.io/csi/csi-hostpath-provisioning-4462^e0b2c93a-e802-11eb-b14a-a6ffd9dc2bdc\") from node \"10.116.0.5\" \nI0718 20:01:04.373858       1 event.go:291] \"Event occurred\" object=\"provisioning-4462/hostpath-client\" kind=\"Pod\" apiVersion=\"v1\" type=\"Normal\" reason=\"SuccessfulAttachVolume\" message=\"AttachVolume.Attach succeeded for volume \\\"pvc-78fc2fe1-4514-4aa4-9fcf-c451cccbfb5a\\\" \"\nI0718 20:01:04.506134       1 namespace_controller.go:185] Namespace has been deleted persistent-local-volumes-test-7373\nE0718 20:01:05.896979       1 reflector.go:138] k8s.io/client-go/metadata/metadatainformer/informer.go:90: Failed to watch *v1.PartialObjectMetadata: failed to list *v1.PartialObjectMetadata: the server could not find the requested resource\nI0718 20:01:06.179523       1 pvc_protection_controller.go:291] \"PVC is unused\" PVC=\"provisioning-1938/pvc-n5sq9\"\nI0718 20:01:06.193391       1 pv_controller.go:640] volume \"local-5jc7w\" is released and reclaim policy \"Retain\" will be executed\nI0718 20:01:06.202840       1 pv_controller.go:879] volume \"local-5jc7w\" entered phase \"Released\"\nI0718 20:01:06.223980       1 pv_controller_base.go:505] deletion of claim \"provisioning-1938/pvc-n5sq9\" was already processed\nI0718 20:01:06.228762       1 namespace_controller.go:185] Namespace has been deleted container-probe-2268\nI0718 20:01:07.275630       1 namespace_controller.go:185] Namespace has been deleted disruption-2-1142\nI0718 20:01:07.333023       1 namespace_controller.go:185] Namespace has been deleted disruption-667\nI0718 20:01:07.720428       1 pv_controller.go:879] volume \"local-pvzw8xs\" entered phase \"Available\"\nI0718 20:01:07.748744       1 pv_controller.go:930] claim \"persistent-local-volumes-test-3617/pvc-x88n2\" bound to volume \"local-pvzw8xs\"\nI0718 20:01:07.760541       1 pv_controller.go:879] volume \"local-pvzw8xs\" entered phase \"Bound\"\nI0718 20:01:07.760603       1 pv_controller.go:982] volume \"local-pvzw8xs\" bound to claim \"persistent-local-volumes-test-3617/pvc-x88n2\"\nI0718 20:01:07.773990       1 pv_controller.go:823] claim \"persistent-local-volumes-test-3617/pvc-x88n2\" entered phase \"Bound\"\nE0718 20:01:07.960101       1 namespace_controller.go:162] deletion of namespace services-8331 failed: unexpected items still remain in namespace: services-8331 for gvr: /v1, Resource=pods\nI0718 20:01:08.064141       1 event.go:291] \"Event occurred\" object=\"csi-mock-volumes-1666-4090/csi-mockplugin\" kind=\"StatefulSet\" apiVersion=\"apps/v1\" type=\"Normal\" reason=\"SuccessfulCreate\" message=\"create Pod csi-mockplugin-0 in StatefulSet csi-mockplugin successful\"\nI0718 20:01:08.096596       1 event.go:291] \"Event occurred\" object=\"csi-mock-volumes-1666-4090/csi-mockplugin-attacher\" kind=\"StatefulSet\" apiVersion=\"apps/v1\" type=\"Normal\" reason=\"SuccessfulCreate\" message=\"create Pod csi-mockplugin-attacher-0 in StatefulSet csi-mockplugin-attacher successful\"\nI0718 20:01:08.133311       1 event.go:291] \"Event occurred\" object=\"csi-mock-volumes-1666-4090/csi-mockplugin-resizer\" kind=\"StatefulSet\" apiVersion=\"apps/v1\" type=\"Normal\" reason=\"SuccessfulCreate\" message=\"create Pod csi-mockplugin-resizer-0 in StatefulSet csi-mockplugin-resizer successful\"\nE0718 20:01:08.590233       1 reflector.go:138] k8s.io/client-go/metadata/metadatainformer/informer.go:90: Failed to watch *v1.PartialObjectMetadata: failed to list *v1.PartialObjectMetadata: the server could not find the requested resource\nI0718 20:01:08.726649       1 pvc_protection_controller.go:303] \"Pod uses PVC\" pod=\"persistent-local-volumes-test-2161/pod-d505b156-40f3-46d0-87e9-5ad02a2a5740\" PVC=\"persistent-local-volumes-test-2161/pvc-fdn29\"\nI0718 20:01:08.726670       1 pvc_protection_controller.go:181] \"Keeping PVC because it is being used\" PVC=\"persistent-local-volumes-test-2161/pvc-fdn29\"\nE0718 20:01:09.140796       1 reflector.go:138] k8s.io/client-go/metadata/metadatainformer/informer.go:90: Failed to watch *v1.PartialObjectMetadata: failed to list *v1.PartialObjectMetadata: the server could not find the requested resource\nE0718 20:01:09.689194       1 tokens_controller.go:262] error synchronizing serviceaccount nettest-3720/default: secrets \"default-token-zslht\" is forbidden: unable to create new content in namespace nettest-3720 because it is being terminated\nI0718 20:01:10.339526       1 event.go:291] \"Event occurred\" object=\"volume-expand-5318-6478/csi-hostpath-attacher\" kind=\"StatefulSet\" apiVersion=\"apps/v1\" type=\"Normal\" reason=\"SuccessfulCreate\" message=\"create Pod csi-hostpath-attacher-0 in StatefulSet csi-hostpath-attacher successful\"\nI0718 20:01:10.464083       1 event.go:291] \"Event occurred\" object=\"volume-expand-5318-6478/csi-hostpathplugin\" kind=\"StatefulSet\" apiVersion=\"apps/v1\" type=\"Normal\" reason=\"SuccessfulCreate\" message=\"create Pod csi-hostpathplugin-0 in StatefulSet csi-hostpathplugin successful\"\nI0718 20:01:10.546172       1 event.go:291] \"Event occurred\" object=\"volume-expand-5318-6478/csi-hostpath-provisioner\" kind=\"StatefulSet\" apiVersion=\"apps/v1\" type=\"Normal\" reason=\"SuccessfulCreate\" message=\"create Pod csi-hostpath-provisioner-0 in StatefulSet csi-hostpath-provisioner successful\"\nI0718 20:01:10.622521       1 event.go:291] \"Event occurred\" object=\"volume-expand-5318-6478/csi-hostpath-resizer\" kind=\"StatefulSet\" apiVersion=\"apps/v1\" type=\"Normal\" reason=\"SuccessfulCreate\" message=\"create Pod csi-hostpath-resizer-0 in StatefulSet csi-hostpath-resizer successful\"\nI0718 20:01:10.697509       1 event.go:291] \"Event occurred\" object=\"volume-expand-5318-6478/csi-hostpath-snapshotter\" kind=\"StatefulSet\" apiVersion=\"apps/v1\" type=\"Normal\" reason=\"SuccessfulCreate\" message=\"create Pod csi-hostpath-snapshotter-0 in StatefulSet csi-hostpath-snapshotter successful\"\nI0718 20:01:10.795388       1 event.go:291] \"Event occurred\" object=\"volume-expand-5318/csi-hostpathw425n\" kind=\"PersistentVolumeClaim\" apiVersion=\"v1\" type=\"Normal\" reason=\"ExternalProvisioning\" message=\"waiting for a volume to be created, either by external provisioner \\\"csi-hostpath-volume-expand-5318\\\" or manually created by system administrator\"\nE0718 20:01:11.944248       1 reflector.go:138] k8s.io/client-go/metadata/metadatainformer/informer.go:90: Failed to watch *v1.PartialObjectMetadata: failed to list *v1.PartialObjectMetadata: the server could not find the requested resource\nI0718 20:01:13.000549       1 deployment_controller.go:583] \"Deployment has been deleted\" deployment=\"deployment-4026/test-cleanup-deployment\"\nI0718 20:01:14.224058       1 pvc_protection_controller.go:291] \"PVC is unused\" PVC=\"persistent-local-volumes-test-2161/pvc-fdn29\"\nI0718 20:01:14.239221       1 pv_controller.go:640] volume \"local-pvp4h9c\" is released and reclaim policy \"Retain\" will be executed\nI0718 20:01:14.246463       1 pv_controller.go:879] volume \"local-pvp4h9c\" entered phase \"Released\"\nI0718 20:01:14.254351       1 pv_controller_base.go:505] deletion of claim \"persistent-local-volumes-test-2161/pvc-fdn29\" was already processed\nI0718 20:01:15.863503       1 pv_controller.go:879] volume \"pvc-13c07603-84d3-48c3-8b62-aac91fe4c41a\" entered phase \"Bound\"\nI0718 20:01:15.863537       1 pv_controller.go:982] volume \"pvc-13c07603-84d3-48c3-8b62-aac91fe4c41a\" bound to claim \"volume-expand-5318/csi-hostpathw425n\"\nI0718 20:01:15.883384       1 pv_controller.go:823] claim \"volume-expand-5318/csi-hostpathw425n\" entered phase \"Bound\"\nI0718 20:01:16.333446       1 replica_set.go:559] \"Too few replicas\" replicaSet=\"kubectl-1946/agnhost-primary\" need=1 creating=1\nI0718 20:01:16.344347       1 event.go:291] \"Event occurred\" object=\"kubectl-1946/agnhost-primary\" kind=\"ReplicationController\" apiVersion=\"v1\" type=\"Normal\" reason=\"SuccessfulCreate\" message=\"Created pod: agnhost-primary-gbrpv\"\nI0718 20:01:16.938112       1 namespace_controller.go:185] Namespace has been deleted provisioning-1938\nI0718 20:01:17.000635       1 deployment_controller.go:583] \"Deployment has been deleted\" deployment=\"webhook-8983/sample-webhook-deployment\"\nI0718 20:01:17.074952       1 namespace_controller.go:185] Namespace has been deleted ephemeral-6983-2204\nI0718 20:01:17.334256       1 pv_controller.go:879] volume \"local-pvhsrbk\" entered phase \"Available\"\nI0718 20:01:17.338308       1 pv_controller.go:930] claim \"persistent-local-volumes-test-3306/pvc-wfxdk\" bound to volume \"local-pvhsrbk\"\nI0718 20:01:17.350833       1 pv_controller.go:879] volume \"local-pvhsrbk\" entered phase \"Bound\"\nI0718 20:01:17.350863       1 pv_controller.go:982] volume \"local-pvhsrbk\" bound to claim \"persistent-local-volumes-test-3306/pvc-wfxdk\"\nI0718 20:01:17.363525       1 pv_controller.go:823] claim \"persistent-local-volumes-test-3306/pvc-wfxdk\" entered phase \"Bound\"\nI0718 20:01:17.497998       1 pvc_protection_controller.go:291] \"PVC is unused\" PVC=\"persistent-local-volumes-test-3306/pvc-wfxdk\"\nI0718 20:01:17.513976       1 pv_controller.go:640] volume \"local-pvhsrbk\" is released and reclaim policy \"Retain\" will be executed\nI0718 20:01:17.519502       1 pv_controller.go:879] volume \"local-pvhsrbk\" entered phase \"Released\"\nI0718 20:01:17.540671       1 pv_controller_base.go:505] deletion of claim \"persistent-local-volumes-test-3306/pvc-wfxdk\" was already processed\nI0718 20:01:17.837192       1 event.go:291] \"Event occurred\" object=\"csi-mock-volumes-1666/pvc-wwxmn\" kind=\"PersistentVolumeClaim\" apiVersion=\"v1\" type=\"Normal\" reason=\"ExternalProvisioning\" message=\"waiting for a volume to be created, either by external provisioner \\\"csi-mock-csi-mock-volumes-1666\\\" or manually created by system administrator\"\nI0718 20:01:17.837224       1 event.go:291] \"Event occurred\" object=\"csi-mock-volumes-1666/pvc-wwxmn\" kind=\"PersistentVolumeClaim\" apiVersion=\"v1\" type=\"Normal\" reason=\"ExternalProvisioning\" message=\"waiting for a volume to be created, either by external provisioner \\\"csi-mock-csi-mock-volumes-1666\\\" or manually created by system administrator\"\nI0718 20:01:17.852314       1 pv_controller.go:879] volume \"pvc-993b4bd2-db01-4763-bc16-a8ef4e91f391\" entered phase \"Bound\"\nI0718 20:01:17.852352       1 pv_controller.go:982] volume \"pvc-993b4bd2-db01-4763-bc16-a8ef4e91f391\" bound to claim \"csi-mock-volumes-1666/pvc-wwxmn\"\nI0718 20:01:17.866623       1 pv_controller.go:823] claim \"csi-mock-volumes-1666/pvc-wwxmn\" entered phase \"Bound\"\nE0718 20:01:17.986564       1 pv_controller.go:1452] error finding provisioning plugin for claim ephemeral-599/inline-volume-46ws4-my-volume: storageclass.storage.k8s.io \"no-such-storage-class\" not found\nI0718 20:01:17.986798       1 event.go:291] \"Event occurred\" object=\"ephemeral-599/inline-volume-46ws4-my-volume\" kind=\"PersistentVolumeClaim\" apiVersion=\"v1\" type=\"Warning\" reason=\"ProvisioningFailed\" message=\"storageclass.storage.k8s.io \\\"no-such-storage-class\\\" not found\"\nI0718 20:01:18.073447       1 graph_builder.go:587] add [v1/Pod, namespace: ephemeral-599, name: inline-volume-46ws4, uid: 46419f7e-e0c9-43be-b32c-f1d70acdd2f3] to the attemptToDelete, because it's waiting for its dependents to be deleted\nI0718 20:01:18.073507       1 garbagecollector.go:471] \"Processing object\" object=\"ephemeral-599/inline-volume-46ws4-my-volume\" objectUID=69453fb2-b723-44ad-b52c-3640007f02c5 kind=\"PersistentVolumeClaim\" virtual=false\nI0718 20:01:18.073645       1 garbagecollector.go:471] \"Processing object\" object=\"ephemeral-599/inline-volume-46ws4\" objectUID=46419f7e-e0c9-43be-b32c-f1d70acdd2f3 kind=\"Pod\" virtual=false\nI0718 20:01:18.085331       1 garbagecollector.go:595] adding [v1/PersistentVolumeClaim, namespace: ephemeral-599, name: inline-volume-46ws4-my-volume, uid: 69453fb2-b723-44ad-b52c-3640007f02c5] to attemptToDelete, because its owner [v1/Pod, namespace: ephemeral-599, name: inline-volume-46ws4, uid: 46419f7e-e0c9-43be-b32c-f1d70acdd2f3] is deletingDependents\nI0718 20:01:18.087422       1 garbagecollector.go:580] \"Deleting object\" object=\"ephemeral-599/inline-volume-46ws4-my-volume\" objectUID=69453fb2-b723-44ad-b52c-3640007f02c5 kind=\"PersistentVolumeClaim\" propagationPolicy=Background\nE0718 20:01:18.093777       1 pv_controller.go:1452] error finding provisioning plugin for claim ephemeral-599/inline-volume-46ws4-my-volume: storageclass.storage.k8s.io \"no-such-storage-class\" not found\nI0718 20:01:18.093806       1 event.go:291] \"Event occurred\" object=\"ephemeral-599/inline-volume-46ws4-my-volume\" kind=\"PersistentVolumeClaim\" apiVersion=\"v1\" type=\"Warning\" reason=\"ProvisioningFailed\" message=\"storageclass.storage.k8s.io \\\"no-such-storage-class\\\" not found\"\nI0718 20:01:18.094155       1 garbagecollector.go:471] \"Processing object\" object=\"ephemeral-599/inline-volume-46ws4-my-volume\" objectUID=69453fb2-b723-44ad-b52c-3640007f02c5 kind=\"PersistentVolumeClaim\" virtual=false\nI0718 20:01:18.096839       1 pvc_protection_controller.go:291] \"PVC is unused\" PVC=\"ephemeral-599/inline-volume-46ws4-my-volume\"\nI0718 20:01:18.102471       1 garbagecollector.go:471] \"Processing object\" object=\"ephemeral-599/inline-volume-46ws4\" objectUID=46419f7e-e0c9-43be-b32c-f1d70acdd2f3 kind=\"Pod\" virtual=false\nI0718 20:01:18.104631       1 garbagecollector.go:590] remove DeleteDependents finalizer for item [v1/Pod, namespace: ephemeral-599, name: inline-volume-46ws4, uid: 46419f7e-e0c9-43be-b32c-f1d70acdd2f3]\nI0718 20:01:19.436961       1 namespace_controller.go:185] Namespace has been deleted persistent-local-volumes-test-2161\nE0718 20:01:19.599966       1 reflector.go:138] k8s.io/client-go/metadata/metadatainformer/informer.go:90: Failed to watch *v1.PartialObjectMetadata: failed to list *v1.PartialObjectMetadata: the server could not find the requested resource\nI0718 20:01:20.016336       1 reconciler.go:295] attacherDetacher.AttachVolume started for volume \"pvc-993b4bd2-db01-4763-bc16-a8ef4e91f391\" (UniqueName: \"kubernetes.io/csi/csi-mock-csi-mock-volumes-1666^4\") from node \"10.116.0.7\" \nI0718 20:01:20.533471       1 namespace_controller.go:185] Namespace has been deleted provisioning-8707\nI0718 20:01:20.580693       1 operation_generator.go:368] AttachVolume.Attach succeeded for volume \"pvc-993b4bd2-db01-4763-bc16-a8ef4e91f391\" (UniqueName: \"kubernetes.io/csi/csi-mock-csi-mock-volumes-1666^4\") from node \"10.116.0.7\" \nI0718 20:01:20.580790       1 event.go:291] \"Event occurred\" object=\"csi-mock-volumes-1666/pvc-volume-tester-4829p\" kind=\"Pod\" apiVersion=\"v1\" type=\"Normal\" reason=\"SuccessfulAttachVolume\" message=\"AttachVolume.Attach succeeded for volume \\\"pvc-993b4bd2-db01-4763-bc16-a8ef4e91f391\\\" \"\nI0718 20:01:21.129164       1 event.go:291] \"Event occurred\" object=\"ephemeral-599-8917/csi-hostpath-attacher\" kind=\"StatefulSet\" apiVersion=\"apps/v1\" type=\"Normal\" reason=\"SuccessfulCreate\" message=\"create Pod csi-hostpath-attacher-0 in StatefulSet csi-hostpath-attacher successful\"\nI0718 20:01:21.238106       1 event.go:291] \"Event occurred\" object=\"ephemeral-599-8917/csi-hostpathplugin\" kind=\"StatefulSet\" apiVersion=\"apps/v1\" type=\"Normal\" reason=\"SuccessfulCreate\" message=\"create Pod csi-hostpathplugin-0 in StatefulSet csi-hostpathplugin successful\"\nI0718 20:01:21.308492       1 event.go:291] \"Event occurred\" object=\"ephemeral-599-8917/csi-hostpath-provisioner\" kind=\"StatefulSet\" apiVersion=\"apps/v1\" type=\"Normal\" reason=\"SuccessfulCreate\" message=\"create Pod csi-hostpath-provisioner-0 in StatefulSet csi-hostpath-provisioner successful\"\nI0718 20:01:21.379863       1 event.go:291] \"Event occurred\" object=\"ephemeral-599-8917/csi-hostpath-resizer\" kind=\"StatefulSet\" apiVersion=\"apps/v1\" type=\"Normal\" reason=\"SuccessfulCreate\" message=\"create Pod csi-hostpath-resizer-0 in StatefulSet csi-hostpath-resizer successful\"\nI0718 20:01:21.739061       1 event.go:291] \"Event occurred\" object=\"ephemeral-599-8917/csi-hostpath-snapshotter\" kind=\"StatefulSet\" apiVersion=\"apps/v1\" type=\"Normal\" reason=\"SuccessfulCreate\" message=\"create Pod csi-hostpath-snapshotter-0 in StatefulSet csi-hostpath-snapshotter successful\"\nI0718 20:01:21.825295       1 event.go:291] \"Event occurred\" object=\"ephemeral-599/inline-volume-tester-fg8zx-my-volume-0\" kind=\"PersistentVolumeClaim\" apiVersion=\"v1\" type=\"Normal\" reason=\"WaitForPodScheduled\" message=\"waiting for pod inline-volume-tester-fg8zx to be scheduled\"\nE0718 20:01:22.489079       1 tokens_controller.go:262] error synchronizing serviceaccount container-probe-4572/default: secrets \"default-token-988gh\" is forbidden: unable to create new content in namespace container-probe-4572 because it is being terminated\nI0718 20:01:23.240625       1 namespace_controller.go:185] Namespace has been deleted ephemeral-8120-5728\nI0718 20:01:23.800089       1 event.go:291] \"Event occurred\" object=\"ephemeral-599/inline-volume-tester-fg8zx-my-volume-0\" kind=\"PersistentVolumeClaim\" apiVersion=\"v1\" type=\"Normal\" reason=\"ExternalProvisioning\" message=\"waiting for a volume to be created, either by external provisioner \\\"csi-hostpath-ephemeral-599\\\" or manually created by system administrator\"\nE0718 20:01:24.706874       1 namespace_controller.go:162] deletion of namespace disruption-5763 failed: unexpected items still remain in namespace: disruption-5763 for gvr: /v1, Resource=pods\nE0718 20:01:24.889824       1 namespace_controller.go:162] deletion of namespace disruption-5763 failed: unexpected items still remain in namespace: disruption-5763 for gvr: /v1, Resource=pods\nE0718 20:01:25.024368       1 namespace_controller.go:162] deletion of namespace apply-7798 failed: unexpected items still remain in namespace: apply-7798 for gvr: /v1, Resource=pods\nE0718 20:01:25.059546       1 namespace_controller.go:162] deletion of namespace disruption-5763 failed: unexpected items still remain in namespace: disruption-5763 for gvr: /v1, Resource=pods\nE0718 20:01:25.187332       1 namespace_controller.go:162] deletion of namespace apply-7798 failed: unexpected items still remain in namespace: apply-7798 for gvr: /v1, Resource=pods\nE0718 20:01:25.251493       1 namespace_controller.go:162] deletion of namespace disruption-5763 failed: unexpected items still remain in namespace: disruption-5763 for gvr: /v1, Resource=pods\nE0718 20:01:25.348338       1 namespace_controller.go:162] deletion of namespace apply-7798 failed: unexpected items still remain in namespace: apply-7798 for gvr: /v1, Resource=pods\nE0718 20:01:25.454920       1 namespace_controller.go:162] deletion of namespace disruption-5763 failed: unexpected items still remain in namespace: disruption-5763 for gvr: /v1, Resource=pods\nE0718 20:01:25.525758       1 namespace_controller.go:162] deletion of namespace apply-7798 failed: unexpected items still remain in namespace: apply-7798 for gvr: /v1, Resource=pods\nE0718 20:01:25.709122       1 namespace_controller.go:162] deletion of namespace disruption-5763 failed: unexpected items still remain in namespace: disruption-5763 for gvr: /v1, Resource=pods\nE0718 20:01:25.717414       1 namespace_controller.go:162] deletion of namespace apply-7798 failed: unexpected items still remain in namespace: apply-7798 for gvr: /v1, Resource=pods\nE0718 20:01:25.977672       1 namespace_controller.go:162] deletion of namespace apply-7798 failed: unexpected items still remain in namespace: apply-7798 for gvr: /v1, Resource=pods\nI0718 20:01:26.001112       1 deployment_controller.go:583] \"Deployment has been deleted\" deployment=\"webhook-7425/sample-webhook-deployment\"\nE0718 20:01:26.048051       1 namespace_controller.go:162] deletion of namespace disruption-5763 failed: unexpected items still remain in namespace: disruption-5763 for gvr: /v1, Resource=pods\nE0718 20:01:26.302108       1 namespace_controller.go:162] deletion of namespace apply-7798 failed: unexpected items still remain in namespace: apply-7798 for gvr: /v1, Resource=pods\nE0718 20:01:26.552704       1 namespace_controller.go:162] deletion of namespace disruption-5763 failed: unexpected items still remain in namespace: disruption-5763 for gvr: /v1, Resource=pods\nE0718 20:01:26.774004       1 namespace_controller.go:162] deletion of namespace apply-7798 failed: unexpected items still remain in namespace: apply-7798 for gvr: /v1, Resource=pods\nE0718 20:01:27.362320       1 namespace_controller.go:162] deletion of namespace disruption-5763 failed: unexpected items still remain in namespace: disruption-5763 for gvr: /v1, Resource=pods\nE0718 20:01:27.558922       1 namespace_controller.go:162] deletion of namespace apply-7798 failed: unexpected items still remain in namespace: apply-7798 for gvr: /v1, Resource=pods\nI0718 20:01:27.678208       1 namespace_controller.go:185] Namespace has been deleted container-probe-4572\nI0718 20:01:27.994107       1 namespace_controller.go:185] Namespace has been deleted tables-1799\nI0718 20:01:28.159576       1 namespace_controller.go:185] Namespace has been deleted persistent-local-volumes-test-3306\nE0718 20:01:28.221403       1 reflector.go:138] k8s.io/client-go/metadata/metadatainformer/informer.go:90: Failed to watch *v1.PartialObjectMetadata: failed to list *v1.PartialObjectMetadata: the server could not find the requested resource\nI0718 20:01:28.663839       1 pvc_protection_controller.go:291] \"PVC is unused\" PVC=\"pv-3628/pvc-wkzch\"\nI0718 20:01:28.674581       1 pv_controller.go:640] volume \"nfs-68cf7\" is released and reclaim policy \"Retain\" will be executed\nI0718 20:01:28.680554       1 pv_controller.go:879] volume \"nfs-68cf7\" entered phase \"Released\"\nI0718 20:01:28.795853       1 pv_controller_base.go:505] deletion of claim \"pv-3628/pvc-wkzch\" was already processed\nE0718 20:01:28.819497       1 namespace_controller.go:162] deletion of namespace disruption-5763 failed: unexpected items still remain in namespace: disruption-5763 for gvr: /v1, Resource=pods\nE0718 20:01:29.000061       1 namespace_controller.go:162] deletion of namespace apply-7798 failed: unexpected items still remain in namespace: apply-7798 for gvr: /v1, Resource=pods\nE0718 20:01:29.616855       1 tokens_controller.go:262] error synchronizing serviceaccount pods-7614/default: secrets \"default-token-qwkvh\" is forbidden: unable to create new content in namespace pods-7614 because it is being terminated\nI0718 20:01:30.993532       1 pv_controller.go:879] volume \"pvc-5ac74b95-2112-475c-9568-a2c5e6e5ef2b\" entered phase \"Bound\"\nI0718 20:01:30.993566       1 pv_controller.go:982] volume \"pvc-5ac74b95-2112-475c-9568-a2c5e6e5ef2b\" bound to claim \"ephemeral-599/inline-volume-tester-fg8zx-my-volume-0\"\nI0718 20:01:31.002760       1 pv_controller.go:823] claim \"ephemeral-599/inline-volume-tester-fg8zx-my-volume-0\" entered phase \"Bound\"\nE0718 20:01:31.350561       1 reflector.go:138] k8s.io/client-go/metadata/metadatainformer/informer.go:90: Failed to watch *v1.PartialObjectMetadata: failed to list *v1.PartialObjectMetadata: the server could not find the requested resource\nI0718 20:01:31.410752       1 namespace_controller.go:185] Namespace has been deleted configmap-1304\nE0718 20:01:31.721647       1 namespace_controller.go:162] deletion of namespace apply-7798 failed: unexpected items still remain in namespace: apply-7798 for gvr: /v1, Resource=pods\nI0718 20:01:31.902488       1 reconciler.go:295] attacherDetacher.AttachVolume started for volume \"pvc-5ac74b95-2112-475c-9568-a2c5e6e5ef2b\" (UniqueName: \"kubernetes.io/csi/csi-hostpath-ephemeral-599^f1d72751-e802-11eb-8a90-6ecc6361dcfa\") from node \"10.116.0.7\" \nI0718 20:01:32.479801       1 operation_generator.go:368] AttachVolume.Attach succeeded for volume \"pvc-5ac74b95-2112-475c-9568-a2c5e6e5ef2b\" (UniqueName: \"kubernetes.io/csi/csi-hostpath-ephemeral-599^f1d72751-e802-11eb-8a90-6ecc6361dcfa\") from node \"10.116.0.7\" \nI0718 20:01:32.479959       1 event.go:291] \"Event occurred\" object=\"ephemeral-599/inline-volume-tester-fg8zx\" kind=\"Pod\" apiVersion=\"v1\" type=\"Normal\" reason=\"SuccessfulAttachVolume\" message=\"AttachVolume.Attach succeeded for volume \\\"pvc-5ac74b95-2112-475c-9568-a2c5e6e5ef2b\\\" \"\nE0718 20:01:32.922031       1 pv_controller.go:1452] error finding provisioning plugin for claim volume-6255/pvc-4xrbd: storageclass.storage.k8s.io \"volume-6255\" not found\nI0718 20:01:32.922085       1 event.go:291] \"Event occurred\" object=\"volume-6255/pvc-4xrbd\" kind=\"PersistentVolumeClaim\" apiVersion=\"v1\" type=\"Warning\" reason=\"ProvisioningFailed\" message=\"storageclass.storage.k8s.io \\\"volume-6255\\\" not found\"\nI0718 20:01:32.961371       1 pv_controller.go:879] volume \"local-9cz82\" entered phase \"Available\"\nI0718 20:01:33.420572       1 graph_builder.go:587] add [v1/Pod, namespace: ephemeral-5827, name: inline-volume-tester2-qw2vr, uid: fee9b4b0-5efd-4b4e-8164-bfc700bc2c63] to the attemptToDelete, because it's waiting for its dependents to be deleted\nI0718 20:01:33.420777       1 garbagecollector.go:471] \"Processing object\" object=\"ephemeral-5827/inline-volume-tester2-qw2vr\" objectUID=fee9b4b0-5efd-4b4e-8164-bfc700bc2c63 kind=\"Pod\" virtual=false\nI0718 20:01:33.420995       1 garbagecollector.go:471] \"Processing object\" object=\"ephemeral-5827/inline-volume-tester2-qw2vr-my-volume-0\" objectUID=cd083b17-8ca9-40cf-80e6-aefca09f6ef7 kind=\"PersistentVolumeClaim\" virtual=false\nI0718 20:01:33.425088       1 garbagecollector.go:595] adding [v1/PersistentVolumeClaim, namespace: ephemeral-5827, name: inline-volume-tester2-qw2vr-my-volume-0, uid: cd083b17-8ca9-40cf-80e6-aefca09f6ef7] to attemptToDelete, because its owner [v1/Pod, namespace: ephemeral-5827, name: inline-volume-tester2-qw2vr, uid: fee9b4b0-5efd-4b4e-8164-bfc700bc2c63] is deletingDependents\nI0718 20:01:33.426309       1 garbagecollector.go:580] \"Deleting object\" object=\"ephemeral-5827/inline-volume-tester2-qw2vr-my-volume-0\" objectUID=cd083b17-8ca9-40cf-80e6-aefca09f6ef7 kind=\"PersistentVolumeClaim\" propagationPolicy=Background\nI0718 20:01:33.432155       1 garbagecollector.go:471] \"Processing object\" object=\"ephemeral-5827/inline-volume-tester2-qw2vr-my-volume-0\" objectUID=cd083b17-8ca9-40cf-80e6-aefca09f6ef7 kind=\"PersistentVolumeClaim\" virtual=false\nI0718 20:01:33.432407       1 pvc_protection_controller.go:303] \"Pod uses PVC\" pod=\"ephemeral-5827/inline-volume-tester2-qw2vr\" PVC=\"ephemeral-5827/inline-volume-tester2-qw2vr-my-volume-0\"\nI0718 20:01:33.432424       1 pvc_protection_controller.go:181] \"Keeping PVC because it is being used\" PVC=\"ephemeral-5827/inline-volume-tester2-qw2vr-my-volume-0\"\nI0718 20:01:33.436462       1 garbagecollector.go:580] \"Deleting object\" object=\"ephemeral-5827/inline-volume-tester2-qw2vr-my-volume-0\" objectUID=cd083b17-8ca9-40cf-80e6-aefca09f6ef7 kind=\"PersistentVolumeClaim\" propagationPolicy=Background\nE0718 20:01:34.716817       1 reflector.go:138] k8s.io/client-go/metadata/metadatainformer/informer.go:90: Failed to watch *v1.PartialObjectMetadata: failed to list *v1.PartialObjectMetadata: the server could not find the requested resource\nI0718 20:01:34.723923       1 namespace_controller.go:185] Namespace has been deleted pods-7614\nE0718 20:01:35.584681       1 reflector.go:138] k8s.io/client-go/metadata/metadatainformer/informer.go:90: Failed to watch *v1.PartialObjectMetadata: failed to list *v1.PartialObjectMetadata: the server could not find the requested resource\nI0718 20:01:36.083838       1 namespace_controller.go:185] Namespace has been deleted nettest-3720\nE0718 20:01:36.359686       1 reflector.go:138] k8s.io/client-go/metadata/metadatainformer/informer.go:90: Failed to watch *v1.PartialObjectMetadata: failed to list *v1.PartialObjectMetadata: the server could not find the requested resource\nI0718 20:01:36.541859       1 namespace_controller.go:185] Namespace has been deleted disruption-5763\nI0718 20:01:36.568148       1 pv_controller.go:930] claim \"volume-6255/pvc-4xrbd\" bound to volume \"local-9cz82\"\nI0718 20:01:36.581378       1 pv_controller.go:879] volume \"local-9cz82\" entered phase \"Bound\"\nI0718 20:01:36.581401       1 pv_controller.go:982] volume \"local-9cz82\" bound to claim \"volume-6255/pvc-4xrbd\"\nI0718 20:01:36.598711       1 pv_controller.go:823] claim \"volume-6255/pvc-4xrbd\" entered phase \"Bound\"\nI0718 20:01:37.753146       1 garbagecollector.go:471] \"Processing object\" object=\"ephemeral-7768-2370/csi-hostpath-attacher-vblj8\" objectUID=4f10d8a8-03a5-48e5-acc5-5cbaa1b63ca8 kind=\"EndpointSlice\" virtual=false\nI0718 20:01:37.758856       1 garbagecollector.go:580] \"Deleting object\" object=\"ephemeral-7768-2370/csi-hostpath-attacher-vblj8\" objectUID=4f10d8a8-03a5-48e5-acc5-5cbaa1b63ca8 kind=\"EndpointSlice\" propagationPolicy=Background\nI0718 20:01:37.799849       1 garbagecollector.go:471] \"Processing object\" object=\"ephemeral-7768-2370/csi-hostpath-attacher-bfbdd46d8\" objectUID=354e18db-1324-4a8b-a077-3fa23a849782 kind=\"ControllerRevision\" virtual=false\nI0718 20:01:37.799897       1 stateful_set.go:419] StatefulSet has been deleted ephemeral-7768-2370/csi-hostpath-attacher\nI0718 20:01:37.799944       1 garbagecollector.go:471] \"Processing object\" object=\"ephemeral-7768-2370/csi-hostpath-attacher-0\" objectUID=2127b07c-d4d5-4b8c-8b10-712ef1a6db21 kind=\"Pod\" virtual=false\nI0718 20:01:37.802496       1 garbagecollector.go:580] \"Deleting object\" object=\"ephemeral-7768-2370/csi-hostpath-attacher-bfbdd46d8\" objectUID=354e18db-1324-4a8b-a077-3fa23a849782 kind=\"ControllerRevision\" propagationPolicy=Background\nI0718 20:01:37.803415       1 garbagecollector.go:580] \"Deleting object\" object=\"ephemeral-7768-2370/csi-hostpath-attacher-0\" objectUID=2127b07c-d4d5-4b8c-8b10-712ef1a6db21 kind=\"Pod\" propagationPolicy=Background\nI0718 20:01:37.870092       1 garbagecollector.go:471] \"Processing object\" object=\"ephemeral-7768-2370/csi-hostpathplugin-vz7cg\" objectUID=a363181d-b084-46eb-84e5-1eb5b0b1d29e kind=\"EndpointSlice\" virtual=false\nI0718 20:01:37.872341       1 garbagecollector.go:580] \"Deleting object\" object=\"ephemeral-7768-2370/csi-hostpathplugin-vz7cg\" objectUID=a363181d-b084-46eb-84e5-1eb5b0b1d29e kind=\"EndpointSlice\" propagationPolicy=Background\nI0718 20:01:37.920959       1 garbagecollector.go:471] \"Processing object\" object=\"ephemeral-7768-2370/csi-hostpathplugin-6566dc7d6\" objectUID=dc69a937-8e9d-4e46-8f35-b9c62adb453b kind=\"ControllerRevision\" virtual=false\nI0718 20:01:37.921035       1 garbagecollector.go:471] \"Processing object\" object=\"ephemeral-7768-2370/csi-hostpathplugin-0\" objectUID=0fc6a403-e9e4-4339-b86c-fdb10eb86063 kind=\"Pod\" virtual=false\nI0718 20:01:37.921097       1 stateful_set.go:419] StatefulSet has been deleted ephemeral-7768-2370/csi-hostpathplugin\nI0718 20:01:37.923547       1 garbagecollector.go:580] \"Deleting object\" object=\"ephemeral-7768-2370/csi-hostpathplugin-6566dc7d6\" objectUID=dc69a937-8e9d-4e46-8f35-b9c62adb453b kind=\"ControllerRevision\" propagationPolicy=Background\nI0718 20:01:37.923670       1 garbagecollector.go:580] \"Deleting object\" object=\"ephemeral-7768-2370/csi-hostpathplugin-0\" objectUID=0fc6a403-e9e4-4339-b86c-fdb10eb86063 kind=\"Pod\" propagationPolicy=Background\nI0718 20:01:37.960149       1 garbagecollector.go:471] \"Processing object\" object=\"ephemeral-7768-2370/csi-hostpath-provisioner-t9f2k\" objectUID=3eee86fc-105b-468a-9a8a-e98f3ddffa7d kind=\"EndpointSlice\" virtual=false\nI0718 20:01:37.962637       1 garbagecollector.go:580] \"Deleting object\" object=\"ephemeral-7768-2370/csi-hostpath-provisioner-t9f2k\" objectUID=3eee86fc-105b-468a-9a8a-e98f3ddffa7d kind=\"EndpointSlice\" propagationPolicy=Background\nI0718 20:01:38.010715       1 garbagecollector.go:471] \"Processing object\" object=\"ephemeral-7768-2370/csi-hostpath-provisioner-6c6dd967cd\" objectUID=5906cc1c-f832-4681-8120-5fd08e42e0b8 kind=\"ControllerRevision\" virtual=false\nI0718 20:01:38.010737       1 garbagecollector.go:471] \"Processing object\" object=\"ephemeral-7768-2370/csi-hostpath-provisioner-0\" objectUID=3cd85eb0-a82e-4c3b-b81a-f5606630848d kind=\"Pod\" virtual=false\nI0718 20:01:38.010715       1 stateful_set.go:419] StatefulSet has been deleted ephemeral-7768-2370/csi-hostpath-provisioner\nI0718 20:01:38.019678       1 garbagecollector.go:580] \"Deleting object\" object=\"ephemeral-7768-2370/csi-hostpath-provisioner-0\" objectUID=3cd85eb0-a82e-4c3b-b81a-f5606630848d kind=\"Pod\" propagationPolicy=Background\nI0718 20:01:38.020324       1 garbagecollector.go:580] \"Deleting object\" object=\"ephemeral-7768-2370/csi-hostpath-provisioner-6c6dd967cd\" objectUID=5906cc1c-f832-4681-8120-5fd08e42e0b8 kind=\"ControllerRevision\" propagationPolicy=Background\nI0718 20:01:38.043682       1 garbagecollector.go:471] \"Processing object\" object=\"ephemeral-7768-2370/csi-hostpath-resizer-68jjl\" objectUID=a112e629-9f64-49ac-9694-cc11214c56ac kind=\"EndpointSlice\" virtual=false\nI0718 20:01:38.046065       1 garbagecollector.go:580] \"Deleting object\" object=\"ephemeral-7768-2370/csi-hostpath-resizer-68jjl\" objectUID=a112e629-9f64-49ac-9694-cc11214c56ac kind=\"EndpointSlice\" propagationPolicy=Background\nI0718 20:01:38.098395       1 garbagecollector.go:471] \"Processing object\" object=\"ephemeral-7768-2370/csi-hostpath-resizer-58f9b79f55\" objectUID=eafaccc0-6bdb-4078-9764-01eea9be81ab kind=\"ControllerRevision\" virtual=false\nI0718 20:01:38.098417       1 stateful_set.go:419] StatefulSet has been deleted ephemeral-7768-2370/csi-hostpath-resizer\nI0718 20:01:38.098528       1 garbagecollector.go:471] \"Processing object\" object=\"ephemeral-7768-2370/csi-hostpath-resizer-0\" objectUID=f8a9d32c-bd26-4224-a1e4-9bd5e92e6c4e kind=\"Pod\" virtual=false\nI0718 20:01:38.100824       1 garbagecollector.go:580] \"Deleting object\" object=\"ephemeral-7768-2370/csi-hostpath-resizer-58f9b79f55\" objectUID=eafaccc0-6bdb-4078-9764-01eea9be81ab kind=\"ControllerRevision\" propagationPolicy=Background\nI0718 20:01:38.101745       1 garbagecollector.go:580] \"Deleting object\" object=\"ephemeral-7768-2370/csi-hostpath-resizer-0\" objectUID=f8a9d32c-bd26-4224-a1e4-9bd5e92e6c4e kind=\"Pod\" propagationPolicy=Background\nI0718 20:01:38.134542       1 garbagecollector.go:471] \"Processing object\" object=\"ephemeral-7768-2370/csi-hostpath-snapshotter-cn48f\" objectUID=6b560c84-f0ee-42cf-9167-8f23b3b58d7a kind=\"EndpointSlice\" virtual=false\nI0718 20:01:38.148908       1 garbagecollector.go:580] \"Deleting object\" object=\"ephemeral-7768-2370/csi-hostpath-snapshotter-cn48f\" objectUID=6b560c84-f0ee-42cf-9167-8f23b3b58d7a kind=\"EndpointSlice\" propagationPolicy=Background\nI0718 20:01:38.194345       1 garbagecollector.go:471] \"Processing object\" object=\"ephemeral-7768-2370/csi-hostpath-snapshotter-64566dfb75\" objectUID=7814690f-2e8b-4396-ad76-603aebb364fd kind=\"ControllerRevision\" virtual=false\nI0718 20:01:38.194423       1 garbagecollector.go:471] \"Processing object\" object=\"ephemeral-7768-2370/csi-hostpath-snapshotter-0\" objectUID=59ca5235-47f2-4b10-b807-459ee471b486 kind=\"Pod\" virtual=false\nI0718 20:01:38.194349       1 stateful_set.go:419] StatefulSet has been deleted ephemeral-7768-2370/csi-hostpath-snapshotter\nI0718 20:01:38.196562       1 garbagecollector.go:580] \"Deleting object\" object=\"ephemeral-7768-2370/csi-hostpath-snapshotter-64566dfb75\" objectUID=7814690f-2e8b-4396-ad76-603aebb364fd kind=\"ControllerRevision\" propagationPolicy=Background\nI0718 20:01:38.197302       1 garbagecollector.go:580] \"Deleting object\" object=\"ephemeral-7768-2370/csi-hostpath-snapshotter-0\" objectUID=59ca5235-47f2-4b10-b807-459ee471b486 kind=\"Pod\" propagationPolicy=Background\nI0718 20:01:38.559041       1 reconciler.go:219] attacherDetacher.DetachVolume started for volume \"pvc-16ff8267-3caf-4d13-92fb-0d8c5cf3bed0\" (UniqueName: \"kubernetes.io/csi/csi-mock-csi-mock-volumes-735^4\") on node \"10.116.0.7\" \nI0718 20:01:38.561770       1 operation_generator.go:1483] Verified volume is safe to detach for volume \"pvc-16ff8267-3caf-4d13-92fb-0d8c5cf3bed0\" (UniqueName: \"kubernetes.io/csi/csi-mock-csi-mock-volumes-735^4\") on node \"10.116.0.7\" \nI0718 20:01:38.903429       1 pvc_protection_controller.go:291] \"PVC is unused\" PVC=\"csi-mock-volumes-5970/pvc-pw64q\"\nI0718 20:01:38.916219       1 pv_controller.go:640] volume \"pvc-471640f7-8ba5-4c6d-a059-e4a0fe35948f\" is released and reclaim policy \"Delete\" will be executed\nI0718 20:01:38.922564       1 pv_controller.go:879] volume \"pvc-471640f7-8ba5-4c6d-a059-e4a0fe35948f\" entered phase \"Released\"\nI0718 20:01:38.924742       1 pv_controller.go:1341] isVolumeReleased[pvc-471640f7-8ba5-4c6d-a059-e4a0fe35948f]: volume is released\nI0718 20:01:38.938892       1 pv_controller_base.go:505] deletion of claim \"csi-mock-volumes-5970/pvc-pw64q\" was already processed\nI0718 20:01:39.127994       1 operation_generator.go:483] DetachVolume.Detach succeeded for volume \"pvc-16ff8267-3caf-4d13-92fb-0d8c5cf3bed0\" (UniqueName: \"kubernetes.io/csi/csi-mock-csi-mock-volumes-735^4\") on node \"10.116.0.7\" \nI0718 20:01:39.322225       1 namespace_controller.go:185] Namespace has been deleted subpath-8921\nI0718 20:01:39.879114       1 reconciler.go:219] attacherDetacher.DetachVolume started for volume \"pvc-919825ea-7889-48f9-9d20-64716ca37948\" (UniqueName: \"kubernetes.io/csi/csi-hostpath-provisioning-1806^de487053-e802-11eb-afb0-3e205f8df9a8\") on node \"10.116.0.5\" \nI0718 20:01:39.882457       1 operation_generator.go:1483] Verified volume is safe to detach for volume \"pvc-919825ea-7889-48f9-9d20-64716ca37948\" (UniqueName: \"kubernetes.io/csi/csi-hostpath-provisioning-1806^de487053-e802-11eb-afb0-3e205f8df9a8\") on node \"10.116.0.5\" \nI0718 20:01:40.132238       1 expand_controller.go:289] Ignoring the PVC \"csi-mock-volumes-1666/pvc-wwxmn\" (uid: \"993b4bd2-db01-4763-bc16-a8ef4e91f391\") : didn't find a plugin capable of expanding the volume; waiting for an external controller to process this PVC.\nI0718 20:01:40.132357       1 event.go:291] \"Event occurred\" object=\"csi-mock-volumes-1666/pvc-wwxmn\" kind=\"PersistentVolumeClaim\" apiVersion=\"v1\" type=\"Warning\" reason=\"ExternalExpanding\" message=\"Ignoring the PVC: didn't find a plugin capable of expanding the volume; waiting for an external controller to process this PVC.\"\nI0718 20:01:40.334360       1 pvc_protection_controller.go:291] \"PVC is unused\" PVC=\"csi-mock-volumes-735/pvc-nwzkz\"\nI0718 20:01:40.349097       1 pv_controller.go:640] volume \"pvc-16ff8267-3caf-4d13-92fb-0d8c5cf3bed0\" is released and reclaim policy \"Delete\" will be executed\nI0718 20:01:40.355757       1 pv_controller.go:879] volume \"pvc-16ff8267-3caf-4d13-92fb-0d8c5cf3bed0\" entered phase \"Released\"\nI0718 20:01:40.358107       1 pv_controller.go:1341] isVolumeReleased[pvc-16ff8267-3caf-4d13-92fb-0d8c5cf3bed0]: volume is released\nI0718 20:01:40.382308       1 pv_controller_base.go:505] deletion of claim \"csi-mock-volumes-735/pvc-nwzkz\" was already processed\nI0718 20:01:40.425819       1 operation_generator.go:483] DetachVolume.Detach succeeded for volume \"pvc-919825ea-7889-48f9-9d20-64716ca37948\" (UniqueName: \"kubernetes.io/csi/csi-hostpath-provisioning-1806^de487053-e802-11eb-afb0-3e205f8df9a8\") on node \"10.116.0.5\" \nE0718 20:01:41.121593       1 reflector.go:138] k8s.io/client-go/metadata/metadatainformer/informer.go:90: Failed to watch *v1.PartialObjectMetadata: failed to list *v1.PartialObjectMetadata: the server could not find the requested resource\nI0718 20:01:41.210197       1 namespace_controller.go:185] Namespace has been deleted ephemeral-7768\nI0718 20:01:41.267009       1 pvc_protection_controller.go:291] \"PVC is unused\" PVC=\"provisioning-1806/csi-hostpath4ndks\"\nI0718 20:01:41.282213       1 pv_controller.go:640] volume \"pvc-919825ea-7889-48f9-9d20-64716ca37948\" is released and reclaim policy \"Delete\" will be executed\nI0718 20:01:41.287939       1 pv_controller.go:879] volume \"pvc-919825ea-7889-48f9-9d20-64716ca37948\" entered phase \"Released\"\nI0718 20:01:41.290841       1 pv_controller.go:1341] isVolumeReleased[pvc-919825ea-7889-48f9-9d20-64716ca37948]: volume is released\nI0718 20:01:41.313341       1 pv_controller_base.go:505] deletion of claim \"provisioning-1806/csi-hostpath4ndks\" was already processed\nI0718 20:01:41.868751       1 pvc_protection_controller.go:303] \"Pod uses PVC\" pod=\"persistent-local-volumes-test-3617/pod-1bcb7fb7-fe6a-4c57-ad87-0a1a8e6c9070\" PVC=\"persistent-local-volumes-test-3617/pvc-x88n2\"\nI0718 20:01:41.868947       1 pvc_protection_controller.go:181] \"Keeping PVC because it is being used\" PVC=\"persistent-local-volumes-test-3617/pvc-x88n2\"\nI0718 20:01:42.000705       1 deployment_controller.go:583] \"Deployment has been deleted\" deployment=\"deployment-9538/webserver-deployment\"\nI0718 20:01:42.027380       1 namespace_controller.go:185] Namespace has been deleted apply-7798\nE0718 20:01:42.032153       1 reflector.go:138] k8s.io/client-go/metadata/metadatainformer/informer.go:90: Failed to watch *v1.PartialObjectMetadata: failed to list *v1.PartialObjectMetadata: the server could not find the requested resource\nI0718 20:01:43.100684       1 namespace_controller.go:185] Namespace has been deleted cronjob-8504\nE0718 20:01:43.685516       1 namespace_controller.go:162] deletion of namespace ephemeral-7768-2370 failed: unable to retrieve the complete list of server APIs: kubectl.example.com/v1: the server could not find the requested resource\nE0718 20:01:43.720697       1 reflector.go:138] k8s.io/client-go/metadata/metadatainformer/informer.go:90: Failed to watch *v1.PartialObjectMetadata: failed to list *v1.PartialObjectMetadata: the server could not find the requested resource\nE0718 20:01:43.849964       1 namespace_controller.go:162] deletion of namespace ephemeral-7768-2370 failed: unable to retrieve the complete list of server APIs: kubectl.example.com/v1: the server could not find the requested resource\nE0718 20:01:44.023362       1 namespace_controller.go:162] deletion of namespace ephemeral-7768-2370 failed: unable to retrieve the complete list of server APIs: kubectl.example.com/v1: the server could not find the requested resource\nE0718 20:01:44.194511       1 namespace_controller.go:162] deletion of namespace ephemeral-7768-2370 failed: unable to retrieve the complete list of server APIs: kubectl.example.com/v1: the server could not find the requested resource\nE0718 20:01:44.403999       1 namespace_controller.go:162] deletion of namespace ephemeral-7768-2370 failed: unable to retrieve the complete list of server APIs: kubectl.example.com/v1: the server could not find the requested resource\nE0718 20:01:44.682182       1 namespace_controller.go:162] deletion of namespace ephemeral-7768-2370 failed: unable to retrieve the complete list of server APIs: kubectl.example.com/v1: the server could not find the requested resource\nE0718 20:01:44.771484       1 pv_controller.go:1452] error finding provisioning plugin for claim provisioning-9310/pvc-49j46: storageclass.storage.k8s.io \"provisioning-9310\" not found\nI0718 20:01:44.771530       1 event.go:291] \"Event occurred\" object=\"provisioning-9310/pvc-49j46\" kind=\"PersistentVolumeClaim\" apiVersion=\"v1\" type=\"Warning\" reason=\"ProvisioningFailed\" message=\"storageclass.storage.k8s.io \\\"provisioning-9310\\\" not found\"\nI0718 20:01:44.811788       1 pv_controller.go:879] volume \"local-5tr89\" entered phase \"Available\"\nE0718 20:01:45.021704       1 namespace_controller.go:162] deletion of namespace ephemeral-7768-2370 failed: unable to retrieve the complete list of server APIs: kubectl.example.com/v1: the server could not find the requested resource\nE0718 20:01:45.394241       1 reflector.go:138] k8s.io/client-go/metadata/metadatainformer/informer.go:90: Failed to watch *v1.PartialObjectMetadata: failed to list *v1.PartialObjectMetadata: the server could not find the requested resource\nE0718 20:01:45.523158       1 namespace_controller.go:162] deletion of namespace ephemeral-7768-2370 failed: unable to retrieve the complete list of server APIs: kubectl.example.com/v1: the server could not find the requested resource\nE0718 20:01:46.186115       1 tokens_controller.go:262] error synchronizing serviceaccount csi-mock-volumes-5970/default: secrets \"default-token-gjspb\" is forbidden: unable to create new content in namespace csi-mock-volumes-5970 because it is being terminated\nE0718 20:01:46.268114       1 namespace_controller.go:162] deletion of namespace pv-3628 failed: unable to retrieve the complete list of server APIs: kubectl.example.com/v1: the server could not find the requested resource\nE0718 20:01:46.314483       1 namespace_controller.go:162] deletion of namespace csi-mock-volumes-5970 failed: unable to retrieve the complete list of server APIs: kubectl.example.com/v1: the server could not find the requested resource\nI0718 20:01:46.362551       1 pvc_protection_controller.go:291] \"PVC is unused\" PVC=\"csi-mock-volumes-1666/pvc-wwxmn\"\nI0718 20:01:46.376408       1 pv_controller.go:640] volume \"pvc-993b4bd2-db01-4763-bc16-a8ef4e91f391\" is released and reclaim policy \"Delete\" will be executed\nI0718 20:01:46.382368       1 pv_controller.go:879] volume \"pvc-993b4bd2-db01-4763-bc16-a8ef4e91f391\" entered phase \"Released\"\nI0718 20:01:46.385071       1 pv_controller.go:1341] isVolumeReleased[pvc-993b4bd2-db01-4763-bc16-a8ef4e91f391]: volume is released\nE0718 20:01:46.394385       1 namespace_controller.go:162] deletion of namespace ephemeral-7768-2370 failed: unable to retrieve the complete list of server APIs: kubectl.example.com/v1: the server could not find the requested resource\nE0718 20:01:46.462281       1 namespace_controller.go:162] deletion of namespace pv-3628 failed: unable to retrieve the complete list of server APIs: kubectl.example.com/v1: the server could not find the requested resource\nI0718 20:01:46.463228       1 namespace_controller.go:185] Namespace has been deleted dns-1104\nE0718 20:01:46.491180       1 namespace_controller.go:162] deletion of namespace csi-mock-volumes-5970 failed: unable to retrieve the complete list of server APIs: kubectl.example.com/v1: the server could not find the requested resource\nE0718 20:01:46.629143       1 namespace_controller.go:162] deletion of namespace pv-3628 failed: unable to retrieve the complete list of server APIs: kubectl.example.com/v1: the server could not find the requested resource\nE0718 20:01:46.669957       1 namespace_controller.go:162] deletion of namespace csi-mock-volumes-5970 failed: unable to retrieve the complete list of server APIs: kubectl.example.com/v1: the server could not find the requested resource\nE0718 20:01:46.825051       1 namespace_controller.go:162] deletion of namespace pv-3628 failed: unable to retrieve the complete list of server APIs: kubectl.example.com/v1: the server could not find the requested resource\nE0718 20:01:46.869263       1 namespace_controller.go:162] deletion of namespace csi-mock-volumes-5970 failed: unable to retrieve the complete list of server APIs: kubectl.example.com/v1: the server could not find the requested resource\nE0718 20:01:47.026501       1 namespace_controller.go:162] deletion of namespace pv-3628 failed: unable to retrieve the complete list of server APIs: kubectl.example.com/v1: the server could not find the requested resource\nE0718 20:01:47.071121       1 namespace_controller.go:162] deletion of namespace csi-mock-volumes-5970 failed: unable to retrieve the complete list of server APIs: kubectl.example.com/v1: the server could not find the requested resource\nI0718 20:01:47.125657       1 namespace_controller.go:185] Namespace has been deleted init-container-9392\nI0718 20:01:47.208711       1 pvc_protection_controller.go:291] \"PVC is unused\" PVC=\"volume-expand-5318/csi-hostpathw425n\"\nI0718 20:01:47.226841       1 pv_controller.go:640] volume \"pvc-13c07603-84d3-48c3-8b62-aac91fe4c41a\" is released and reclaim policy \"Delete\" will be executed\nI0718 20:01:47.231776       1 pv_controller.go:879] volume \"pvc-13c07603-84d3-48c3-8b62-aac91fe4c41a\" entered phase \"Released\"\nI0718 20:01:47.234062       1 pv_controller.go:1341] isVolumeReleased[pvc-13c07603-84d3-48c3-8b62-aac91fe4c41a]: volume is released\nI0718 20:01:47.257004       1 pv_controller_base.go:505] deletion of claim \"volume-expand-5318/csi-hostpathw425n\" was already processed\nE0718 20:01:47.467145       1 namespace_controller.go:162] deletion of namespace pv-3628 failed: unable to retrieve the complete list of server APIs: kubectl.example.com/v1: the server could not find the requested resource\nE0718 20:01:47.509245       1 namespace_controller.go:162] deletion of namespace csi-mock-volumes-5970 failed: unable to retrieve the complete list of server APIs: kubectl.example.com/v1: the server could not find the requested resource\nE0718 20:01:47.585223       1 tokens_controller.go:262] error synchronizing serviceaccount csi-mock-volumes-735/default: secrets \"default-token-fvfjq\" is forbidden: unable to create new content in namespace csi-mock-volumes-735 because it is being terminated\nE0718 20:01:47.650027       1 tokens_controller.go:262] error synchronizing serviceaccount persistent-local-volumes-test-3617/default: secrets \"default-token-wgvq5\" is forbidden: unable to create new content in namespace persistent-local-volumes-test-3617 because it is being terminated\nI0718 20:01:47.681781       1 pvc_protection_controller.go:303] \"Pod uses PVC\" pod=\"persistent-local-volumes-test-3617/pod-1bcb7fb7-fe6a-4c57-ad87-0a1a8e6c9070\" PVC=\"persistent-local-volumes-test-3617/pvc-x88n2\"\nI0718 20:01:47.681809       1 pvc_protection_controller.go:181] \"Keeping PVC because it is being used\" PVC=\"persistent-local-volumes-test-3617/pvc-x88n2\"\nI0718 20:01:47.702938       1 pvc_protection_controller.go:291] \"PVC is unused\" PVC=\"persistent-local-volumes-test-3617/pvc-x88n2\"\nI0718 20:01:47.716517       1 pv_controller.go:640] volume \"local-pvzw8xs\" is released and reclaim policy \"Retain\" will be executed\nI0718 20:01:47.724324       1 pv_controller.go:879] volume \"local-pvzw8xs\" entered phase \"Released\"\nI0718 20:01:47.735045       1 pv_controller_base.go:505] deletion of claim \"persistent-local-volumes-test-3617/pvc-x88n2\" was already processed\nE0718 20:01:47.830581       1 namespace_controller.go:162] deletion of namespace csi-mock-volumes-735 failed: unable to retrieve the complete list of server APIs: kubectl.example.com/v1: the server could not find the requested resource\nE0718 20:01:47.871217       1 namespace_controller.go:162] deletion of namespace persistent-local-volumes-test-3617 failed: unable to retrieve the complete list of server APIs: kubectl.example.com/v1: the server could not find the requested resource\nW0718 20:01:48.095273       1 utils.go:265] Service services-7615/service-headless using reserved endpoint slices label, skipping label service.kubernetes.io/headless: \nI0718 20:01:48.095308       1 replica_set.go:559] \"Too few replicas\" replicaSet=\"services-7615/service-headless\" need=3 creating=1\nW0718 20:01:48.107277       1 utils.go:265] Service services-7615/service-headless using reserved endpoint slices label, skipping label service.kubernetes.io/headless: \nW0718 20:01:48.117355       1 utils.go:265] Service services-7615/service-headless using reserved endpoint slices label, skipping label service.kubernetes.io/headless: \nI0718 20:01:48.125912       1 replica_set.go:559] \"Too few replicas\" replicaSet=\"services-7615/service-headless-toggled\" need=3 creating=1\nI0718 20:01:48.192224       1 garbagecollector.go:471] \"Processing object\" object=\"services-7615/service-headless-7l89j\" objectUID=2f340c99-d0f8-404e-bf50-27097c5ece2e kind=\"Pod\" virtual=false\nI0718 20:01:48.192258       1 garbagecollector.go:471] \"Processing object\" object=\"services-7615/service-headless-j2p67\" objectUID=39a194ff-c37c-441c-8a97-bf64d2e4f8cd kind=\"Pod\" virtual=false\nI0718 20:01:48.192271       1 garbagecollector.go:471] \"Processing object\" object=\"services-7615/service-headless-ld62n\" objectUID=ff7d3d91-811f-44c8-971f-938bb9664ac5 kind=\"Pod\" virtual=false\nI0718 20:01:48.199378       1 garbagecollector.go:471] \"Processing object\" object=\"services-7615/service-headless-toggled-fsz6g\" objectUID=490f1d64-838d-4a83-ad17-f75e1442d82a kind=\"Pod\" virtual=false\nI0718 20:01:48.199392       1 garbagecollector.go:471] \"Processing object\" object=\"services-7615/service-headless-toggled-7cl4v\" objectUID=4fcbc7aa-22ad-4543-9e10-aeac14d71e1a kind=\"Pod\" virtual=false\nI0718 20:01:48.199406       1 garbagecollector.go:471] \"Processing object\" object=\"services-7615/service-headless-toggled-b2wm6\" objectUID=edaf7fb7-681f-4e32-9b41-5492dbecf8b7 kind=\"Pod\" virtual=false\nI0718 20:01:48.210514       1 garbagecollector.go:471] \"Processing object\" object=\"services-7615/service-headless-kmfbd\" objectUID=7cec8d98-08f5-43af-865c-2b534140488a kind=\"EndpointSlice\" virtual=false\nI0718 20:01:48.213706       1 garbagecollector.go:580] \"Deleting object\" object=\"services-7615/service-headless-kmfbd\" objectUID=7cec8d98-08f5-43af-865c-2b534140488a kind=\"EndpointSlice\" propagationPolicy=Background\nI0718 20:01:48.228547       1 garbagecollector.go:471] \"Processing object\" object=\"services-7615/service-headless-toggled-4tbb8\" objectUID=79eb9171-7e2f-47e8-af7f-6c007127f94d kind=\"EndpointSlice\" virtual=false\nI0718 20:01:48.230752       1 garbagecollector.go:580] \"Deleting object\" object=\"services-7615/service-headless-toggled-4tbb8\" objectUID=79eb9171-7e2f-47e8-af7f-6c007127f94d kind=\"EndpointSlice\" propagationPolicy=Background\nE0718 20:01:48.283703       1 namespace_controller.go:162] deletion of namespace services-7615 failed: [unable to retrieve the complete list of server APIs: kubectl.example.com/v1: the server could not find the requested resource, unexpected items still remain in namespace: services-7615 for gvr: /v1, Resource=pods]\nE0718 20:01:48.356248       1 reflector.go:138] k8s.io/client-go/metadata/metadatainformer/informer.go:90: Failed to watch *v1.PartialObjectMetadata: failed to list *v1.PartialObjectMetadata: the server could not find the requested resource\nE0718 20:01:48.478619       1 namespace_controller.go:162] deletion of namespace services-7615 failed: unexpected items still remain in namespace: services-7615 for gvr: /v1, Resource=pods\nE0718 20:01:48.695703       1 namespace_controller.go:162] deletion of namespace services-7615 failed: unexpected items still remain in namespace: services-7615 for gvr: /v1, Resource=pods\nE0718 20:01:48.897005       1 namespace_controller.go:162] deletion of namespace services-7615 failed: unexpected items still remain in namespace: services-7615 for gvr: /v1, Resource=pods\nE0718 20:01:49.161299       1 namespace_controller.go:162] deletion of namespace services-7615 failed: unexpected items still remain in namespace: services-7615 for gvr: /v1, Resource=pods\nE0718 20:01:49.441278       1 namespace_controller.go:162] deletion of namespace services-7615 failed: unexpected items still remain in namespace: services-7615 for gvr: /v1, Resource=pods\nI0718 20:01:49.558181       1 stateful_set.go:419] StatefulSet has been deleted csi-mock-volumes-735-8114/csi-mockplugin\nI0718 20:01:49.558232       1 garbagecollector.go:471] \"Processing object\" object=\"csi-mock-volumes-735-8114/csi-mockplugin-78d6577b77\" objectUID=fc0e4076-d939-4d1b-9816-4f1da5de8149 kind=\"ControllerRevision\" virtual=false\nI0718 20:01:49.558236       1 garbagecollector.go:471] \"Processing object\" object=\"csi-mock-volumes-735-8114/csi-mockplugin-0\" objectUID=d477102c-3c3c-4be3-8476-16d1af5868ea kind=\"Pod\" virtual=false\nI0718 20:01:49.560624       1 garbagecollector.go:580] \"Deleting object\" object=\"csi-mock-volumes-735-8114/csi-mockplugin-78d6577b77\" objectUID=fc0e4076-d939-4d1b-9816-4f1da5de8149 kind=\"ControllerRevision\" propagationPolicy=Background\nI0718 20:01:49.561528       1 garbagecollector.go:580] \"Deleting object\" object=\"csi-mock-volumes-735-8114/csi-mockplugin-0\" objectUID=d477102c-3c3c-4be3-8476-16d1af5868ea kind=\"Pod\" propagationPolicy=Background\nI0718 20:01:49.627109       1 garbagecollector.go:471] \"Processing object\" object=\"csi-mock-volumes-735-8114/csi-mockplugin-attacher-879cd4797\" objectUID=9cb30334-e7c5-456a-96e6-3e4c90308d81 kind=\"ControllerRevision\" virtual=false\nI0718 20:01:49.627169       1 stateful_set.go:419] StatefulSet has been deleted csi-mock-volumes-735-8114/csi-mockplugin-attacher\nI0718 20:01:49.627292       1 garbagecollector.go:471] \"Processing object\" object=\"csi-mock-volumes-735-8114/csi-mockplugin-attacher-0\" objectUID=8d53c445-50e1-4685-9818-e40258c840c5 kind=\"Pod\" virtual=false\nI0718 20:01:49.630121       1 garbagecollector.go:580] \"Deleting object\" object=\"csi-mock-volumes-735-8114/csi-mockplugin-attacher-879cd4797\" objectUID=9cb30334-e7c5-456a-96e6-3e4c90308d81 kind=\"ControllerRevision\" propagationPolicy=Background\nI0718 20:01:49.630133       1 garbagecollector.go:580] \"Deleting object\" object=\"csi-mock-volumes-735-8114/csi-mockplugin-attacher-0\" objectUID=8d53c445-50e1-4685-9818-e40258c840c5 kind=\"Pod\" propagationPolicy=Background\nI0718 20:01:49.658485       1 replica_set.go:559] \"Too few replicas\" replicaSet=\"kubectl-1946/agnhost-primary\" need=1 creating=1\nI0718 20:01:49.715966       1 garbagecollector.go:471] \"Processing object\" object=\"kubectl-1946/agnhost-primary-gbrpv\" objectUID=e8ff722e-f2ba-4353-857c-99ea2fd16e66 kind=\"Pod\" virtual=false\nE0718 20:01:49.814917       1 namespace_controller.go:162] deletion of namespace services-7615 failed: unexpected items still remain in namespace: services-7615 for gvr: /v1, Resource=pods\nE0718 20:01:49.925206       1 pv_controller.go:1452] error finding provisioning plugin for claim volume-7899/pvc-kxxc4: storageclass.storage.k8s.io \"volume-7899\" not found\nI0718 20:01:49.925362       1 event.go:291] \"Event occurred\" object=\"volume-7899/pvc-kxxc4\" kind=\"PersistentVolumeClaim\" apiVersion=\"v1\" type=\"Warning\" reason=\"ProvisioningFailed\" message=\"storageclass.storage.k8s.io \\\"volume-7899\\\" not found\"\nI0718 20:01:49.966379       1 pv_controller.go:879] volume \"local-vrxvf\" entered phase \"Available\"\nI0718 20:01:50.094038       1 garbagecollector.go:471] \"Processing object\" object=\"csi-mock-volumes-5970-9198/csi-mockplugin-c8f6d54d5\" objectUID=4cd07472-c2bc-4ea8-b2e9-a9af9bcbb812 kind=\"ControllerRevision\" virtual=false\nI0718 20:01:50.094054       1 garbagecollector.go:471] \"Processing object\" object=\"csi-mock-volumes-5970-9198/csi-mockplugin-0\" objectUID=7344ae09-65e6-414a-8625-021bd96bb748 kind=\"Pod\" virtual=false\nI0718 20:01:50.094131       1 stateful_set.go:419] StatefulSet has been deleted csi-mock-volumes-5970-9198/csi-mockplugin\nI0718 20:01:50.096462       1 garbagecollector.go:580] \"Deleting object\" object=\"csi-mock-volumes-5970-9198/csi-mockplugin-c8f6d54d5\" objectUID=4cd07472-c2bc-4ea8-b2e9-a9af9bcbb812 kind=\"ControllerRevision\" propagationPolicy=Background\nI0718 20:01:50.098526       1 garbagecollector.go:580] \"Deleting object\" object=\"csi-mock-volumes-5970-9198/csi-mockplugin-0\" objectUID=7344ae09-65e6-414a-8625-021bd96bb748 kind=\"Pod\" propagationPolicy=Background\nE0718 20:01:50.315630       1 namespace_controller.go:162] deletion of namespace services-7615 failed: unexpected items still remain in namespace: services-7615 for gvr: /v1, Resource=pods\nI0718 20:01:50.770902       1 reconciler.go:219] attacherDetacher.DetachVolume started for volume \"pvc-78fc2fe1-4514-4aa4-9fcf-c451cccbfb5a\" (UniqueName: \"kubernetes.io/csi/csi-hostpath-provisioning-4462^e0b2c93a-e802-11eb-b14a-a6ffd9dc2bdc\") on node \"10.116.0.5\" \nI0718 20:01:50.773717       1 operation_generator.go:1483] Verified volume is safe to detach for volume \"pvc-78fc2fe1-4514-4aa4-9fcf-c451cccbfb5a\" (UniqueName: \"kubernetes.io/csi/csi-hostpath-provisioning-4462^e0b2c93a-e802-11eb-b14a-a6ffd9dc2bdc\") on node \"10.116.0.5\" \nE0718 20:01:51.184219       1 namespace_controller.go:162] deletion of namespace services-7615 failed: unexpected items still remain in namespace: services-7615 for gvr: /v1, Resource=pods\nE0718 20:01:51.222441       1 reflector.go:138] k8s.io/client-go/metadata/metadatainformer/informer.go:90: Failed to watch *v1.PartialObjectMetadata: failed to list *v1.PartialObjectMetadata: the server could not find the requested resource\nI0718 20:01:51.335088       1 operation_generator.go:483] DetachVolume.Detach succeeded for volume \"pvc-78fc2fe1-4514-4aa4-9fcf-c451cccbfb5a\" (UniqueName: \"kubernetes.io/csi/csi-hostpath-provisioning-4462^e0b2c93a-e802-11eb-b14a-a6ffd9dc2bdc\") on node \"10.116.0.5\" \nE0718 20:01:51.428414       1 tokens_controller.go:262] error synchronizing serviceaccount provisioning-1806/default: secrets \"default-token-9kntv\" is forbidden: unable to create new content in namespace provisioning-1806 because it is being terminated\nI0718 20:01:51.567843       1 pv_controller.go:930] claim \"volume-7899/pvc-kxxc4\" bound to volume \"local-vrxvf\"\nI0718 20:01:51.574833       1 pv_controller.go:1341] isVolumeReleased[pvc-993b4bd2-db01-4763-bc16-a8ef4e91f391]: volume is released\nI0718 20:01:51.579441       1 pv_controller.go:879] volume \"local-vrxvf\" entered phase \"Bound\"\nI0718 20:01:51.579463       1 pv_controller.go:982] volume \"local-vrxvf\" bound to claim \"volume-7899/pvc-kxxc4\"\nI0718 20:01:51.595311       1 pv_controller.go:823] claim \"volume-7899/pvc-kxxc4\" entered phase \"Bound\"\nI0718 20:01:51.595564       1 pv_controller.go:930] claim \"provisioning-9310/pvc-49j46\" bound to volume \"local-5tr89\"\nI0718 20:01:51.606544       1 pv_controller.go:879] volume \"local-5tr89\" entered phase \"Bound\"\nI0718 20:01:51.606574       1 pv_controller.go:982] volume \"local-5tr89\" bound to claim \"provisioning-9310/pvc-49j46\"\nI0718 20:01:51.618566       1 pv_controller.go:823] claim \"provisioning-9310/pvc-49j46\" entered phase \"Bound\"\nI0718 20:01:52.259879       1 resource_quota_controller.go:435] syncing resource quota controller with updated resources from discovery: added: [kubectl.example.com/v1, Resource=e2e-test-kubectl-4850-crds], removed: []\nI0718 20:01:52.259981       1 resource_quota_monitor.go:229] QuotaMonitor created object count evaluator for e2e-test-kubectl-4850-crds.kubectl.example.com\nI0718 20:01:52.260036       1 shared_informer.go:240] Waiting for caches to sync for resource quota\nI0718 20:01:52.361022       1 shared_informer.go:247] Caches are synced for resource quota \nI0718 20:01:52.361041       1 resource_quota_controller.go:454] synced quota controller\nI0718 20:01:52.491026       1 reconciler.go:219] attacherDetacher.DetachVolume started for volume \"pvc-993b4bd2-db01-4763-bc16-a8ef4e91f391\" (UniqueName: \"kubernetes.io/csi/csi-mock-csi-mock-volumes-1666^4\") on node \"10.116.0.7\" \nI0718 20:01:52.494233       1 operation_generator.go:1483] Verified volume is safe to detach for volume \"pvc-993b4bd2-db01-4763-bc16-a8ef4e91f391\" (UniqueName: \"kubernetes.io/csi/csi-mock-csi-mock-volumes-1666^4\") on node \"10.116.0.7\" \nE0718 20:01:52.649965       1 namespace_controller.go:162] deletion of namespace services-7615 failed: unexpected items still remain in namespace: services-7615 for gvr: /v1, Resource=pods\nI0718 20:01:52.825082       1 garbagecollector.go:213] syncing garbage collector with updated resources from discovery (attempt 1): added: [kubectl.example.com/v1, Resource=e2e-test-kubectl-4850-crds], removed: []\nI0718 20:01:52.831759       1 shared_informer.go:240] Waiting for caches to sync for garbage collector\nI0718 20:01:52.831807       1 shared_informer.go:247] Caches are synced for garbage collector \nI0718 20:01:52.831812       1 garbagecollector.go:254] synced garbage collector\nI0718 20:01:52.910044       1 namespace_controller.go:185] Namespace has been deleted pv-3628\nI0718 20:01:52.934190       1 namespace_controller.go:185] Namespace has been deleted csi-mock-volumes-5970\nI0718 20:01:53.047169       1 operation_generator.go:483] DetachVolume.Detach succeeded for volume \"pvc-993b4bd2-db01-4763-bc16-a8ef4e91f391\" (UniqueName: \"kubernetes.io/csi/csi-mock-csi-mock-volumes-1666^4\") on node \"10.116.0.7\" \nI0718 20:01:53.071194       1 namespace_controller.go:185] Namespace has been deleted csi-mock-volumes-735\nI0718 20:01:53.127579       1 namespace_controller.go:185] Namespace has been deleted persistent-local-volumes-test-3617\nI0718 20:01:53.192812       1 garbagecollector.go:471] \"Processing object\" object=\"provisioning-1806-9608/csi-hostpath-attacher-qvm6f\" objectUID=83dc0b16-e27f-46fc-8d1e-14f0391d733a kind=\"EndpointSlice\" virtual=false\nI0718 20:01:53.196008       1 garbagecollector.go:580] \"Deleting object\" object=\"provisioning-1806-9608/csi-hostpath-attacher-qvm6f\" objectUID=83dc0b16-e27f-46fc-8d1e-14f0391d733a kind=\"EndpointSlice\" propagationPolicy=Background\nI0718 20:01:53.243163       1 stateful_set.go:419] StatefulSet has been deleted provisioning-1806-9608/csi-hostpath-attacher\nI0718 20:01:53.243186       1 garbagecollector.go:471] \"Processing object\" object=\"provisioning-1806-9608/csi-hostpath-attacher-0\" objectUID=6ce52e7a-8512-4ac9-8d56-309318ec74d5 kind=\"Pod\" virtual=false\nI0718 20:01:53.243238       1 garbagecollector.go:471] \"Processing object\" object=\"provisioning-1806-9608/csi-hostpath-attacher-cccbd5dff\" objectUID=e56578d2-fc33-4ff5-bd2f-4bb12ba9609c kind=\"ControllerRevision\" virtual=false\nI0718 20:01:53.246005       1 garbagecollector.go:580] \"Deleting object\" object=\"provisioning-1806-9608/csi-hostpath-attacher-0\" objectUID=6ce52e7a-8512-4ac9-8d56-309318ec74d5 kind=\"Pod\" propagationPolicy=Background\nI0718 20:01:53.246811       1 garbagecollector.go:580] \"Deleting object\" object=\"provisioning-1806-9608/csi-hostpath-attacher-cccbd5dff\" objectUID=e56578d2-fc33-4ff5-bd2f-4bb12ba9609c kind=\"ControllerRevision\" propagationPolicy=Background\nI0718 20:01:53.313585       1 garbagecollector.go:471] \"Processing object\" object=\"provisioning-1806-9608/csi-hostpathplugin-xddmx\" objectUID=0454f534-cd9e-442a-bec6-0f1031e31ea4 kind=\"EndpointSlice\" virtual=false\nI0718 20:01:53.320581       1 garbagecollector.go:580] \"Deleting object\" object=\"provisioning-1806-9608/csi-hostpathplugin-xddmx\" objectUID=0454f534-cd9e-442a-bec6-0f1031e31ea4 kind=\"EndpointSlice\" propagationPolicy=Background\nI0718 20:01:53.362183       1 stateful_set.go:419] StatefulSet has been deleted provisioning-1806-9608/csi-hostpathplugin\nI0718 20:01:53.362183       1 garbagecollector.go:471] \"Processing object\" object=\"provisioning-1806-9608/csi-hostpathplugin-79c67dd7d5\" objectUID=128b68cd-3a2d-473b-9c40-e344756bc457 kind=\"ControllerRevision\" virtual=false\nI0718 20:01:53.362203       1 garbagecollector.go:471] \"Processing object\" object=\"provisioning-1806-9608/csi-hostpathplugin-0\" objectUID=83b64742-dca6-40c8-af37-b17dafb4cf88 kind=\"Pod\" virtual=false\nI0718 20:01:53.364608       1 garbagecollector.go:580] \"Deleting object\" object=\"provisioning-1806-9608/csi-hostpathplugin-79c67dd7d5\" objectUID=128b68cd-3a2d-473b-9c40-e344756bc457 kind=\"ControllerRevision\" propagationPolicy=Background\nI0718 20:01:53.365732       1 garbagecollector.go:580] \"Deleting object\" object=\"provisioning-1806-9608/csi-hostpathplugin-0\" objectUID=83b64742-dca6-40c8-af37-b17dafb4cf88 kind=\"Pod\" propagationPolicy=Background\nI0718 20:01:53.402089       1 garbagecollector.go:471] \"Processing object\" object=\"provisioning-1806-9608/csi-hostpath-provisioner-znlbp\" objectUID=bf078ce0-45c5-4e4f-876f-1772e1f2f896 kind=\"EndpointSlice\" virtual=false\nI0718 20:01:53.409085       1 garbagecollector.go:580] \"Deleting object\" object=\"provisioning-1806-9608/csi-hostpath-provisioner-znlbp\" objectUID=bf078ce0-45c5-4e4f-876f-1772e1f2f896 kind=\"EndpointSlice\" propagationPolicy=Background\nI0718 20:01:53.427211       1 pv_controller_base.go:505] deletion of claim \"csi-mock-volumes-1666/pvc-wwxmn\" was already processed\nI0718 20:01:53.453749       1 garbagecollector.go:471] \"Processing object\" object=\"provisioning-1806-9608/csi-hostpath-provisioner-56f9bd77b5\" objectUID=bc0d68cb-a110-4f7d-b46f-30eb413e2016 kind=\"ControllerRevision\" virtual=false\nI0718 20:01:53.453895       1 stateful_set.go:419] StatefulSet has been deleted provisioning-1806-9608/csi-hostpath-provisioner\nI0718 20:01:53.453906       1 garbagecollector.go:471] \"Processing object\" object=\"provisioning-1806-9608/csi-hostpath-provisioner-0\" objectUID=d71f18a5-61ad-491d-8c6b-af16956cf7b9 kind=\"Pod\" virtual=false\nI0718 20:01:53.456062       1 garbagecollector.go:580] \"Deleting object\" object=\"provisioning-1806-9608/csi-hostpath-provisioner-56f9bd77b5\" objectUID=bc0d68cb-a110-4f7d-b46f-30eb413e2016 kind=\"ControllerRevision\" propagationPolicy=Background\nI0718 20:01:53.457011       1 garbagecollector.go:580] \"Deleting object\" object=\"provisioning-1806-9608/csi-hostpath-provisioner-0\" objectUID=d71f18a5-61ad-491d-8c6b-af16956cf7b9 kind=\"Pod\" propagationPolicy=Background\nI0718 20:01:53.488320       1 garbagecollector.go:471] \"Processing object\" object=\"provisioning-1806-9608/csi-hostpath-resizer-njnql\" objectUID=3e4e534e-77cc-4cbf-aa9a-d320f02f40e1 kind=\"EndpointSlice\" virtual=false\nI0718 20:01:53.493599       1 garbagecollector.go:580] \"Deleting object\" object=\"provisioning-1806-9608/csi-hostpath-resizer-njnql\" objectUID=3e4e534e-77cc-4cbf-aa9a-d320f02f40e1 kind=\"EndpointSlice\" propagationPolicy=Background\nI0718 20:01:53.535089       1 garbagecollector.go:471] \"Processing object\" object=\"provisioning-1806-9608/csi-hostpath-resizer-6bb65d96bb\" objectUID=b1fabec7-b58c-4a15-9d0b-0eb03164214e kind=\"ControllerRevision\" virtual=false\nI0718 20:01:53.535106       1 stateful_set.go:419] StatefulSet has been deleted provisioning-1806-9608/csi-hostpath-resizer\nI0718 20:01:53.535118       1 garbagecollector.go:471] \"Processing object\" object=\"provisioning-1806-9608/csi-hostpath-resizer-0\" objectUID=4ebd6217-2eb2-47b8-8bed-e980e3976806 kind=\"Pod\" virtual=false\nI0718 20:01:53.537340       1 garbagecollector.go:580] \"Deleting object\" object=\"provisioning-1806-9608/csi-hostpath-resizer-6bb65d96bb\" objectUID=b1fabec7-b58c-4a15-9d0b-0eb03164214e kind=\"ControllerRevision\" propagationPolicy=Background\nI0718 20:01:53.538415       1 garbagecollector.go:580] \"Deleting object\" object=\"provisioning-1806-9608/csi-hostpath-resizer-0\" objectUID=4ebd6217-2eb2-47b8-8bed-e980e3976806 kind=\"Pod\" propagationPolicy=Background\nI0718 20:01:53.574571       1 garbagecollector.go:471] \"Processing object\" object=\"provisioning-1806-9608/csi-hostpath-snapshotter-r824l\" objectUID=7819c757-2cc9-4021-bfc0-30f45d7a2ea7 kind=\"EndpointSlice\" virtual=false\nI0718 20:01:53.581039       1 garbagecollector.go:580] \"Deleting object\" object=\"provisioning-1806-9608/csi-hostpath-snapshotter-r824l\" objectUID=7819c757-2cc9-4021-bfc0-30f45d7a2ea7 kind=\"EndpointSlice\" propagationPolicy=Background\nI0718 20:01:53.630841       1 stateful_set.go:419] StatefulSet has been deleted provisioning-1806-9608/csi-hostpath-snapshotter\nI0718 20:01:53.630847       1 garbagecollector.go:471] \"Processing object\" object=\"provisioning-1806-9608/csi-hostpath-snapshotter-796785ff4\" objectUID=9a5bc023-1f22-4721-8b59-d911fa8d722b kind=\"ControllerRevision\" virtual=false\nI0718 20:01:53.630849       1 garbagecollector.go:471] \"Processing object\" object=\"provisioning-1806-9608/csi-hostpath-snapshotter-0\" objectUID=a4182137-6601-4664-9916-cf922b398ef3 kind=\"Pod\" virtual=false\nI0718 20:01:53.633084       1 garbagecollector.go:580] \"Deleting object\" object=\"provisioning-1806-9608/csi-hostpath-snapshotter-796785ff4\" objectUID=9a5bc023-1f22-4721-8b59-d911fa8d722b kind=\"ControllerRevision\" propagationPolicy=Background\nI0718 20:01:53.633878       1 garbagecollector.go:580] \"Deleting object\" object=\"provisioning-1806-9608/csi-hostpath-snapshotter-0\" objectUID=a4182137-6601-4664-9916-cf922b398ef3 kind=\"Pod\" propagationPolicy=Background\nI0718 20:01:53.839076       1 pv_controller.go:879] volume \"hostpath-lgcf2\" entered phase \"Available\"\nI0718 20:01:54.666571       1 replica_set.go:559] \"Too few replicas\" replicaSet=\"services-4562/externalsvc\" need=2 creating=2\nI0718 20:01:54.674739       1 event.go:291] \"Event occurred\" object=\"services-4562/externalsvc\" kind=\"ReplicationController\" apiVersion=\"v1\" type=\"Normal\" reason=\"SuccessfulCreate\" message=\"Created pod: externalsvc-vvx8f\"\nI0718 20:01:54.682117       1 event.go:291] \"Event occurred\" object=\"services-4562/externalsvc\" kind=\"ReplicationController\" apiVersion=\"v1\" type=\"Normal\" reason=\"SuccessfulCreate\" message=\"Created pod: externalsvc-ct777\"\nI0718 20:01:55.364115       1 replica_set.go:559] \"Too few replicas\" replicaSet=\"webhook-7902/sample-webhook-deployment-78988fc6cd\" need=1 creating=1\nI0718 20:01:55.364272       1 event.go:291] \"Event occurred\" object=\"webhook-7902/sample-webhook-deployment\" kind=\"Deployment\" apiVersion=\"apps/v1\" type=\"Normal\" reason=\"ScalingReplicaSet\" message=\"Scaled up replica set sample-webhook-deployment-78988fc6cd to 1\"\nI0718 20:01:55.375232       1 event.go:291] \"Event occurred\" object=\"webhook-7902/sample-webhook-deployment-78988fc6cd\" kind=\"ReplicaSet\" apiVersion=\"apps/v1\" type=\"Normal\" reason=\"SuccessfulCreate\" message=\"Created pod: sample-webhook-deployment-78988fc6cd-d4hgx\"\nI0718 20:01:55.376928       1 deployment_controller.go:490] \"Error syncing deployment\" deployment=\"webhook-7902/sample-webhook-deployment\" err=\"Operation cannot be fulfilled on deployments.apps \\\"sample-webhook-deployment\\\": the object has been modified; please apply your changes to the latest version and try again\"\nE0718 20:01:55.403853       1 namespace_controller.go:162] deletion of namespace services-7615 failed: unexpected items still remain in namespace: services-7615 for gvr: /v1, Resource=pods\nI0718 20:01:55.420501       1 pvc_protection_controller.go:291] \"PVC is unused\" PVC=\"provisioning-4462/pvc-bdxmk\"\nI0718 20:01:55.430980       1 pv_controller.go:640] volume \"pvc-78fc2fe1-4514-4aa4-9fcf-c451cccbfb5a\" is released and reclaim policy \"Delete\" will be executed\nI0718 20:01:55.607561       1 namespace_controller.go:185] Namespace has been deleted kubectl-9550\nI0718 20:01:55.686215       1 pv_controller.go:879] volume \"pvc-78fc2fe1-4514-4aa4-9fcf-c451cccbfb5a\" entered phase \"Released\"\nI0718 20:01:55.690588       1 pv_controller.go:1341] isVolumeReleased[pvc-78fc2fe1-4514-4aa4-9fcf-c451cccbfb5a]: volume is released\nI0718 20:01:55.710929       1 pv_controller_base.go:505] deletion of claim \"provisioning-4462/pvc-bdxmk\" was already processed\nI0718 20:01:56.231470       1 namespace_controller.go:185] Namespace has been deleted kubectl-4473\nI0718 20:01:56.276818       1 graph_builder.go:587] add [v1/Pod, namespace: ephemeral-599, name: inline-volume-tester-fg8zx, uid: 121b3ea5-1324-4b58-87cc-328311442635] to the attemptToDelete, because it's waiting for its dependents to be deleted\nI0718 20:01:56.276889       1 garbagecollector.go:471] \"Processing object\" object=\"ephemeral-599/inline-volume-tester-fg8zx-my-volume-0\" objectUID=5ac74b95-2112-475c-9568-a2c5e6e5ef2b kind=\"PersistentVolumeClaim\" virtual=false\nI0718 20:01:56.276928       1 garbagecollector.go:471] \"Processing object\" object=\"ephemeral-599/inline-volume-tester-fg8zx\" objectUID=121b3ea5-1324-4b58-87cc-328311442635 kind=\"Pod\" virtual=false\nI0718 20:01:56.280432       1 garbagecollector.go:595] adding [v1/PersistentVolumeClaim, namespace: ephemeral-599, name: inline-volume-tester-fg8zx-my-volume-0, uid: 5ac74b95-2112-475c-9568-a2c5e6e5ef2b] to attemptToDelete, because its owner [v1/Pod, namespace: ephemeral-599, name: inline-volume-tester-fg8zx, uid: 121b3ea5-1324-4b58-87cc-328311442635] is deletingDependents\nI0718 20:01:56.281611       1 garbagecollector.go:580] \"Deleting object\" object=\"ephemeral-599/inline-volume-tester-fg8zx-my-volume-0\" objectUID=5ac74b95-2112-475c-9568-a2c5e6e5ef2b kind=\"PersistentVolumeClaim\" propagationPolicy=Background\nI0718 20:01:56.288212       1 garbagecollector.go:471] \"Processing object\" object=\"ephemeral-599/inline-volume-tester-fg8zx-my-volume-0\" objectUID=5ac74b95-2112-475c-9568-a2c5e6e5ef2b kind=\"PersistentVolumeClaim\" virtual=false\nI0718 20:01:56.288223       1 pvc_protection_controller.go:303] \"Pod uses PVC\" pod=\"ephemeral-599/inline-volume-tester-fg8zx\" PVC=\"ephemeral-599/inline-volume-tester-fg8zx-my-volume-0\"\nI0718 20:01:56.288242       1 pvc_protection_controller.go:181] \"Keeping PVC because it is being used\" PVC=\"ephemeral-599/inline-volume-tester-fg8zx-my-volume-0\"\nI0718 20:01:56.670042       1 namespace_controller.go:185] Namespace has been deleted provisioning-1806\nE0718 20:01:57.510381       1 tokens_controller.go:262] error synchronizing serviceaccount volume-expand-5318/default: secrets \"default-token-d9jxb\" is forbidden: unable to create new content in namespace volume-expand-5318 because it is being terminated\nI0718 20:01:57.930986       1 pv_controller.go:879] volume \"local-pvf4rsz\" entered phase \"Available\"\nI0718 20:01:57.957740       1 pv_controller.go:930] claim \"persistent-local-volumes-test-7580/pvc-mnzvt\" bound to volume \"local-pvf4rsz\"\nI0718 20:01:57.971540       1 pv_controller.go:879] volume \"local-pvf4rsz\" entered phase \"Bound\"\nI0718 20:01:57.971570       1 pv_controller.go:982] volume \"local-pvf4rsz\" bound to claim \"persistent-local-volumes-test-7580/pvc-mnzvt\"\nI0718 20:01:57.984463       1 pv_controller.go:823] claim \"persistent-local-volumes-test-7580/pvc-mnzvt\" entered phase \"Bound\"\nI0718 20:01:58.096900       1 namespace_controller.go:185] Namespace has been deleted ephemeral-7768-2370\nE0718 20:01:58.844721       1 tokens_controller.go:262] error synchronizing serviceaccount provisioning-1806-9608/default: secrets \"default-token-mvhm7\" is forbidden: unable to create new content in namespace provisioning-1806-9608 because it is being terminated\nI0718 20:01:59.000213       1 deployment_controller.go:583] \"Deployment has been deleted\" deployment=\"apply-6822/deployment\"\nE0718 20:01:59.066696       1 reflector.go:138] k8s.io/client-go/metadata/metadatainformer/informer.go:90: Failed to watch *v1.PartialObjectMetadata: failed to list *v1.PartialObjectMetadata: the server could not find the requested resource\nI0718 20:01:59.240125       1 garbagecollector.go:471] \"Processing object\" object=\"volume-expand-5318-6478/csi-hostpath-attacher-xb4sj\" objectUID=fa213078-2ff8-412a-9020-807a00d362ce kind=\"EndpointSlice\" virtual=false\nI0718 20:01:59.243544       1 garbagecollector.go:580] \"Deleting object\" object=\"volume-expand-5318-6478/csi-hostpath-attacher-xb4sj\" objectUID=fa213078-2ff8-412a-9020-807a00d362ce kind=\"EndpointSlice\" propagationPolicy=Background\nI0718 20:01:59.307896       1 garbagecollector.go:471] \"Processing object\" object=\"volume-expand-5318-6478/csi-hostpath-attacher-7478ffd4b5\" objectUID=e674c600-1042-4952-ad44-18a7f712efb8 kind=\"ControllerRevision\" virtual=false\nI0718 20:01:59.307955       1 stateful_set.go:419] StatefulSet has been deleted volume-expand-5318-6478/csi-hostpath-attacher\nI0718 20:01:59.308001       1 garbagecollector.go:471] \"Processing object\" object=\"volume-expand-5318-6478/csi-hostpath-attacher-0\" objectUID=60f50523-d12b-4b9c-80dc-6a9930ddae89 kind=\"Pod\" virtual=false\nI0718 20:01:59.310505       1 garbagecollector.go:580] \"Deleting object\" object=\"volume-expand-5318-6478/csi-hostpath-attacher-0\" objectUID=60f50523-d12b-4b9c-80dc-6a9930ddae89 kind=\"Pod\" propagationPolicy=Background\nI0718 20:01:59.311115       1 garbagecollector.go:580] \"Deleting object\" object=\"volume-expand-5318-6478/csi-hostpath-attacher-7478ffd4b5\" objectUID=e674c600-1042-4952-ad44-18a7f712efb8 kind=\"ControllerRevision\" propagationPolicy=Background\nI0718 20:01:59.376620       1 garbagecollector.go:471] \"Processing object\" object=\"volume-expand-5318-6478/csi-hostpathplugin-j2r4p\" objectUID=2005d6dc-956e-4757-bed0-42cca65d44d0 kind=\"EndpointSlice\" virtual=false\nI0718 20:01:59.387292       1 garbagecollector.go:580] \"Deleting object\" object=\"volume-expand-5318-6478/csi-hostpathplugin-j2r4p\" objectUID=2005d6dc-956e-4757-bed0-42cca65d44d0 kind=\"EndpointSlice\" propagationPolicy=Background\nE0718 20:01:59.412119       1 tokens_controller.go:262] error synchronizing serviceaccount dns-2451/default: secrets \"default-token-bmttr\" is forbidden: unable to create new content in namespace dns-2451 because it is being terminated\nI0718 20:01:59.434384       1 garbagecollector.go:471] \"Processing object\" object=\"volume-expand-5318-6478/csi-hostpathplugin-76c6b784c5\" objectUID=3c8540dc-95c2-487f-b301-bc00ba681602 kind=\"ControllerRevision\" virtual=false\nI0718 20:01:59.434487       1 stateful_set.go:419] StatefulSet has been deleted volume-expand-5318-6478/csi-hostpathplugin\nI0718 20:01:59.434522       1 garbagecollector.go:471] \"Processing object\" object=\"volume-expand-5318-6478/csi-hostpathplugin-0\" objectUID=f8926a34-c64d-415d-9955-981b8e46ced7 kind=\"Pod\" virtual=false\nI0718 20:01:59.437017       1 garbagecollector.go:580] \"Deleting object\" object=\"volume-expand-5318-6478/csi-hostpathplugin-76c6b784c5\" objectUID=3c8540dc-95c2-487f-b301-bc00ba681602 kind=\"ControllerRevision\" propagationPolicy=Background\nI0718 20:01:59.437111       1 garbagecollector.go:580] \"Deleting object\" object=\"volume-expand-5318-6478/csi-hostpathplugin-0\" objectUID=f8926a34-c64d-415d-9955-981b8e46ced7 kind=\"Pod\" propagationPolicy=Background\nI0718 20:01:59.473038       1 garbagecollector.go:471] \"Processing object\" object=\"volume-expand-5318-6478/csi-hostpath-provisioner-4dh4v\" objectUID=ccd6678b-5c55-4035-a26e-677ea7b589c1 kind=\"EndpointSlice\" virtual=false\nI0718 20:01:59.477548       1 garbagecollector.go:580] \"Deleting object\" object=\"volume-expand-5318-6478/csi-hostpath-provisioner-4dh4v\" objectUID=ccd6678b-5c55-4035-a26e-677ea7b589c1 kind=\"EndpointSlice\" propagationPolicy=Background\nI0718 20:01:59.540706       1 garbagecollector.go:471] \"Processing object\" object=\"volume-expand-5318-6478/csi-hostpath-provisioner-8474fb8b95\" objectUID=4c0a5e28-24ea-4f9a-82a2-c51f0c0e9ebc kind=\"ControllerRevision\" virtual=false\nI0718 20:01:59.540715       1 garbagecollector.go:471] \"Processing object\" object=\"volume-expand-5318-6478/csi-hostpath-provisioner-0\" objectUID=7d9d92c0-6cf9-4e21-8a2d-c40d3329a4d8 kind=\"Pod\" virtual=false\nI0718 20:01:59.540720       1 stateful_set.go:419] StatefulSet has been deleted volume-expand-5318-6478/csi-hostpath-provisioner\nI0718 20:01:59.543084       1 garbagecollector.go:580] \"Deleting object\" object=\"volume-expand-5318-6478/csi-hostpath-provisioner-8474fb8b95\" objectUID=4c0a5e28-24ea-4f9a-82a2-c51f0c0e9ebc kind=\"ControllerRevision\" propagationPolicy=Background\nI0718 20:01:59.543089       1 garbagecollector.go:580] \"Deleting object\" object=\"volume-expand-5318-6478/csi-hostpath-provisioner-0\" objectUID=7d9d92c0-6cf9-4e21-8a2d-c40d3329a4d8 kind=\"Pod\" propagationPolicy=Background\nI0718 20:01:59.580886       1 garbagecollector.go:471] \"Processing object\" object=\"volume-expand-5318-6478/csi-hostpath-resizer-4hdbj\" objectUID=3b7b4401-1c01-4a0f-8623-3bf40e5ac659 kind=\"EndpointSlice\" virtual=false\nI0718 20:01:59.586820       1 garbagecollector.go:580] \"Deleting object\" object=\"volume-expand-5318-6478/csi-hostpath-resizer-4hdbj\" objectUID=3b7b4401-1c01-4a0f-8623-3bf40e5ac659 kind=\"EndpointSlice\" propagationPolicy=Background\nI0718 20:01:59.646600       1 garbagecollector.go:471] \"Processing object\" object=\"volume-expand-5318-6478/csi-hostpath-resizer-6d74b44858\" objectUID=3fe48ac9-a65f-4b8b-940d-7e823596305e kind=\"ControllerRevision\" virtual=false\nI0718 20:01:59.646631       1 stateful_set.go:419] StatefulSet has been deleted volume-expand-5318-6478/csi-hostpath-resizer\nI0718 20:01:59.646642       1 garbagecollector.go:471] \"Processing object\" object=\"volume-expand-5318-6478/csi-hostpath-resizer-0\" objectUID=05ca808d-d52e-4617-8dea-0374b67115ff kind=\"Pod\" virtual=false\nI0718 20:01:59.656685       1 garbagecollector.go:580] \"Deleting object\" object=\"volume-expand-5318-6478/csi-hostpath-resizer-6d74b44858\" objectUID=3fe48ac9-a65f-4b8b-940d-7e823596305e kind=\"ControllerRevision\" propagationPolicy=Background\nI0718 20:01:59.658138       1 garbagecollector.go:580] \"Deleting object\" object=\"volume-expand-5318-6478/csi-hostpath-resizer-0\" objectUID=05ca808d-d52e-4617-8dea-0374b67115ff kind=\"Pod\" propagationPolicy=Background\nE0718 20:01:59.662614       1 tokens_controller.go:262] error synchronizing serviceaccount csi-mock-volumes-1666/default: secrets \"default-token-4g7lc\" is forbidden: unable to create new content in namespace csi-mock-volumes-1666 because it is being terminated\nI0718 20:01:59.689919       1 garbagecollector.go:471] \"Processing object\" object=\"volume-expand-5318-6478/csi-hostpath-snapshotter-9n8pt\" objectUID=478fcd33-b80e-4137-8e63-2e577e5e0fda kind=\"EndpointSlice\" virtual=false\nI0718 20:01:59.692275       1 garbagecollector.go:580] \"Deleting object\" object=\"volume-expand-5318-6478/csi-hostpath-snapshotter-9n8pt\" objectUID=478fcd33-b80e-4137-8e63-2e577e5e0fda kind=\"EndpointSlice\" propagationPolicy=Background\nI0718 20:01:59.759258       1 garbagecollector.go:471] \"Processing object\" object=\"volume-expand-5318-6478/csi-hostpath-snapshotter-67fb997ff4\" objectUID=4bcb1b0c-4ecc-495a-81c2-2a2d1ba240e9 kind=\"ControllerRevision\" virtual=false\nI0718 20:01:59.759293       1 garbagecollector.go:471] \"Processing object\" object=\"volume-expand-5318-6478/csi-hostpath-snapshotter-0\" objectUID=d9c14ad4-62a3-4ccc-8889-7dda6ef903ce kind=\"Pod\" virtual=false\nI0718 20:01:59.759298       1 stateful_set.go:419] StatefulSet has been deleted volume-expand-5318-6478/csi-hostpath-snapshotter\nI0718 20:01:59.761556       1 garbagecollector.go:580] \"Deleting object\" object=\"volume-expand-5318-6478/csi-hostpath-snapshotter-67fb997ff4\" objectUID=4bcb1b0c-4ecc-495a-81c2-2a2d1ba240e9 kind=\"ControllerRevision\" propagationPolicy=Background\nI0718 20:01:59.762353       1 garbagecollector.go:580] \"Deleting object\" object=\"volume-expand-5318-6478/csi-hostpath-snapshotter-0\" objectUID=d9c14ad4-62a3-4ccc-8889-7dda6ef903ce kind=\"Pod\" propagationPolicy=Background\nI0718 20:02:00.120771       1 event.go:291] \"Event occurred\" object=\"cronjob-4272/concurrent\" kind=\"CronJob\" apiVersion=\"batch/v1\" type=\"Normal\" reason=\"SuccessfulCreate\" message=\"Created job concurrent-27110642\"\nI0718 20:02:00.138476       1 event.go:291] \"Event occurred\" object=\"cronjob-4272/concurrent-27110642\" kind=\"Job\" apiVersion=\"batch/v1\" type=\"Normal\" reason=\"SuccessfulCreate\" message=\"Created pod: concurrent-27110642-jzt4h\"\nI0718 20:02:00.140197       1 cronjob_controllerv2.go:193] \"error cleaning up jobs\" cronjob=\"cronjob-4272/concurrent\" resourceVersion=\"30389\" err=\"Operation cannot be fulfilled on cronjobs.batch \\\"concurrent\\\": the object has been modified; please apply your changes to the latest version and try again\"\nE0718 20:02:00.140218       1 cronjob_controllerv2.go:154] error syncing CronJobController cronjob-4272/concurrent, requeuing: Operation cannot be fulfilled on cronjobs.batch \"concurrent\": the object has been modified; please apply your changes to the latest version and try again\nE0718 20:02:00.727935       1 namespace_controller.go:162] deletion of namespace services-7615 failed: unexpected items still remain in namespace: services-7615 for gvr: /v1, Resource=pods\nI0718 20:02:00.806031       1 pvc_protection_controller.go:291] \"PVC is unused\" PVC=\"provisioning-4462/pvc-xv4lq\"\nE0718 20:02:00.822162       1 pvc_protection_controller.go:215] \"Error removing protection finalizer from PVC\" err=\"Operation cannot be fulfilled on persistentvolumeclaims \\\"pvc-xv4lq\\\": the object has been modified; please apply your changes to the latest version and try again\" PVC=\"provisioning-4462/pvc-xv4lq\"\nE0718 20:02:00.822189       1 pvc_protection_controller.go:149] PVC provisioning-4462/pvc-xv4lq failed with : Operation cannot be fulfilled on persistentvolumeclaims \"pvc-xv4lq\": the object has been modified; please apply your changes to the latest version and try again\nI0718 20:02:00.835747       1 pvc_protection_controller.go:291] \"PVC is unused\" PVC=\"provisioning-4462/pvc-xv4lq\"\nI0718 20:02:00.846085       1 pvc_protection_controller.go:291] \"PVC is unused\" PVC=\"provisioning-4462/pvc-xv4lq\"\nI0718 20:02:00.846989       1 pv_controller.go:640] volume \"pvc-1522f266-6376-4263-9b6a-b0d2f01cbf7d\" is released and reclaim policy \"Delete\" will be executed\nE0718 20:02:00.848382       1 pvc_protection_controller.go:215] \"Error removing protection finalizer from PVC\" err=\"Operation cannot be fulfilled on persistentvolumeclaims \\\"pvc-xv4lq\\\": StorageError: invalid object, Code: 4, Key: /registry/persistentvolumeclaims/provisioning-4462/pvc-xv4lq, ResourceVersion: 0, AdditionalErrorMsg: Precondition failed: UID in precondition: 1522f266-6376-4263-9b6a-b0d2f01cbf7d, UID in object meta: \" PVC=\"provisioning-4462/pvc-xv4lq\"\nE0718 20:02:00.848402       1 pvc_protection_controller.go:149] PVC provisioning-4462/pvc-xv4lq failed with : Operation cannot be fulfilled on persistentvolumeclaims \"pvc-xv4lq\": StorageError: invalid object, Code: 4, Key: /registry/persistentvolumeclaims/provisioning-4462/pvc-xv4lq, ResourceVersion: 0, AdditionalErrorMsg: Precondition failed: UID in precondition: 1522f266-6376-4263-9b6a-b0d2f01cbf7d, UID in object meta: \nI0718 20:02:00.854207       1 pv_controller.go:879] volume \"pvc-1522f266-6376-4263-9b6a-b0d2f01cbf7d\" entered phase \"Released\"\nI0718 20:02:00.857006       1 pv_controller.go:1341] isVolumeReleased[pvc-1522f266-6376-4263-9b6a-b0d2f01cbf7d]: volume is released\nI0718 20:02:00.885566       1 pv_controller_base.go:505] deletion of claim \"provisioning-4462/pvc-xv4lq\" was already processed\nE0718 20:02:00.945447       1 reflector.go:138] k8s.io/client-go/metadata/metadatainformer/informer.go:90: Failed to watch *v1.PartialObjectMetadata: the server could not find the requested resource\nE0718 20:02:01.105544       1 pv_controller.go:1452] error finding provisioning plugin for claim volume-6367/pvc-5kgs5: storageclass.storage.k8s.io \"volume-6367\" not found\nI0718 20:02:01.105753       1 event.go:291] \"Event occurred\" object=\"volume-6367/pvc-5kgs5\" kind=\"PersistentVolumeClaim\" apiVersion=\"v1\" type=\"Warning\" reason=\"ProvisioningFailed\" message=\"storageclass.storage.k8s.io \\\"volume-6367\\\" not found\"\nI0718 20:02:01.167158       1 namespace_controller.go:185] Namespace has been deleted security-context-9611\nI0718 20:02:01.178059       1 pv_controller.go:879] volume \"local-m8msb\" entered phase \"Available\"\nE0718 20:02:01.878271       1 reflector.go:138] k8s.io/client-go/metadata/metadatainformer/informer.go:90: Failed to watch *v1.PartialObjectMetadata: failed to list *v1.PartialObjectMetadata: the server could not find the requested resource\nI0718 20:02:01.906053       1 garbagecollector.go:471] \"Processing object\" object=\"csi-mock-volumes-1666-4090/csi-mockplugin-7c997764f7\" objectUID=66f6317e-e423-42a6-9168-d6699b71ed7e kind=\"ControllerRevision\" virtual=false\nI0718 20:02:01.906071       1 garbagecollector.go:471] \"Processing object\" object=\"csi-mock-volumes-1666-4090/csi-mockplugin-0\" objectUID=b2335f9f-b9a2-446a-85f9-9dc212deed1a kind=\"Pod\" virtual=false\nI0718 20:02:01.906075       1 stateful_set.go:419] StatefulSet has been deleted csi-mock-volumes-1666-4090/csi-mockplugin\nI0718 20:02:01.908515       1 garbagecollector.go:580] \"Deleting object\" object=\"csi-mock-volumes-1666-4090/csi-mockplugin-7c997764f7\" objectUID=66f6317e-e423-42a6-9168-d6699b71ed7e kind=\"ControllerRevision\" propagationPolicy=Background\nI0718 20:02:01.909465       1 garbagecollector.go:580] \"Deleting object\" object=\"csi-mock-volumes-1666-4090/csi-mockplugin-0\" objectUID=b2335f9f-b9a2-446a-85f9-9dc212deed1a kind=\"Pod\" propagationPolicy=Background\nI0718 20:02:01.940935       1 garbagecollector.go:471] \"Processing object\" object=\"csi-mock-volumes-1666-4090/csi-mockplugin-attacher-65c8fc7c69\" objectUID=e94ec60c-c42b-4c83-9256-66abf1c9bf0b kind=\"ControllerRevision\" virtual=false\nI0718 20:02:01.940944       1 stateful_set.go:419] StatefulSet has been deleted csi-mock-volumes-1666-4090/csi-mockplugin-attacher\nI0718 20:02:01.940945       1 garbagecollector.go:471] \"Processing object\" object=\"csi-mock-volumes-1666-4090/csi-mockplugin-attacher-0\" objectUID=799106a5-c046-4eb7-bf4e-9183e77c2719 kind=\"Pod\" virtual=false\nI0718 20:02:01.943397       1 garbagecollector.go:580] \"Deleting object\" object=\"csi-mock-volumes-1666-4090/csi-mockplugin-attacher-65c8fc7c69\" objectUID=e94ec60c-c42b-4c83-9256-66abf1c9bf0b kind=\"ControllerRevision\" propagationPolicy=Background\nI0718 20:02:01.944345       1 garbagecollector.go:580] \"Deleting object\" object=\"csi-mock-volumes-1666-4090/csi-mockplugin-attacher-0\" objectUID=799106a5-c046-4eb7-bf4e-9183e77c2719 kind=\"Pod\" propagationPolicy=Background\nI0718 20:02:01.975231       1 garbagecollector.go:471] \"Processing object\" object=\"csi-mock-volumes-1666-4090/csi-mockplugin-resizer-7b699b5b8\" objectUID=4d0c97ac-f934-4ca7-84d1-43142a28ddbc kind=\"ControllerRevision\" virtual=false\nI0718 20:02:01.975243       1 stateful_set.go:419] StatefulSet has been deleted csi-mock-volumes-1666-4090/csi-mockplugin-resizer\nI0718 20:02:01.975299       1 garbagecollector.go:471] \"Processing object\" object=\"csi-mock-volumes-1666-4090/csi-mockplugin-resizer-0\" objectUID=0c93da56-01f0-49d2-b8ce-7b8c7dc010d2 kind=\"Pod\" virtual=false\nI0718 20:02:01.978070       1 garbagecollector.go:580] \"Deleting object\" object=\"csi-mock-volumes-1666-4090/csi-mockplugin-resizer-0\" objectUID=0c93da56-01f0-49d2-b8ce-7b8c7dc010d2 kind=\"Pod\" propagationPolicy=Background\nI0718 20:02:01.978771       1 garbagecollector.go:580] \"Deleting object\" object=\"csi-mock-volumes-1666-4090/csi-mockplugin-resizer-7b699b5b8\" objectUID=4d0c97ac-f934-4ca7-84d1-43142a28ddbc kind=\"ControllerRevision\" propagationPolicy=Background\nI0718 20:02:02.148176       1 garbagecollector.go:471] \"Processing object\" object=\"cronjob-4272/concurrent-27110642\" objectUID=2f08baa7-2aa9-44cb-8d3a-608fb5b1864c kind=\"Job\" virtual=false\nI0718 20:02:02.148179       1 garbagecollector.go:471] \"Processing object\" object=\"cronjob-4272/concurrent-27110641\" objectUID=15013ddf-3480-41b6-bb8e-0a2072828b6e kind=\"Job\" virtual=false\nI0718 20:02:02.151032       1 garbagecollector.go:580] \"Deleting object\" object=\"cronjob-4272/concurrent-27110642\" objectUID=2f08baa7-2aa9-44cb-8d3a-608fb5b1864c kind=\"Job\" propagationPolicy=Background\nI0718 20:02:02.151893       1 garbagecollector.go:580] \"Deleting object\" object=\"cronjob-4272/concurrent-27110641\" objectUID=15013ddf-3480-41b6-bb8e-0a2072828b6e kind=\"Job\" propagationPolicy=Background\nI0718 20:02:02.156497       1 garbagecollector.go:471] \"Processing object\" object=\"cronjob-4272/concurrent-27110642-jzt4h\" objectUID=fa6560d2-2474-4bac-b43a-92c8bda35af1 kind=\"Pod\" virtual=false\nI0718 20:02:02.158330       1 garbagecollector.go:471] \"Processing object\" object=\"cronjob-4272/concurrent-27110641-zx4v2\" objectUID=64566826-7da2-4abd-8c7b-dd1124842b31 kind=\"Pod\" virtual=false\nI0718 20:02:02.158974       1 garbagecollector.go:580] \"Deleting object\" object=\"cronjob-4272/concurrent-27110642-jzt4h\" objectUID=fa6560d2-2474-4bac-b43a-92c8bda35af1 kind=\"Pod\" propagationPolicy=Background\nI0718 20:02:02.160592       1 garbagecollector.go:580] \"Deleting object\" object=\"cronjob-4272/concurrent-27110641-zx4v2\" objectUID=64566826-7da2-4abd-8c7b-dd1124842b31 kind=\"Pod\" propagationPolicy=Background\nI0718 20:02:02.585033       1 namespace_controller.go:185] Namespace has been deleted volume-expand-5318\nE0718 20:02:02.842144       1 tokens_controller.go:262] error synchronizing serviceaccount subpath-8417/default: secrets \"default-token-pl5dm\" is forbidden: unable to create new content in namespace subpath-8417 because it is being terminated\nI0718 20:02:03.045282       1 event.go:291] \"Event occurred\" object=\"provisioning-7043-7873/csi-hostpath-attacher\" kind=\"StatefulSet\" apiVersion=\"apps/v1\" type=\"Normal\" reason=\"SuccessfulCreate\" message=\"create Pod csi-hostpath-attacher-0 in StatefulSet csi-hostpath-attacher successful\"\nI0718 20:02:03.167656       1 event.go:291] \"Event occurred\" object=\"provisioning-7043-7873/csi-hostpathplugin\" kind=\"StatefulSet\" apiVersion=\"apps/v1\" type=\"Normal\" reason=\"SuccessfulCreate\" message=\"create Pod csi-hostpathplugin-0 in StatefulSet csi-hostpathplugin successful\"\nI0718 20:02:03.279219       1 event.go:291] \"Event occurred\" object=\"provisioning-7043-7873/csi-hostpath-provisioner\" kind=\"StatefulSet\" apiVersion=\"apps/v1\" type=\"Normal\" reason=\"SuccessfulCreate\" message=\"create Pod csi-hostpath-provisioner-0 in StatefulSet csi-hostpath-provisioner successful\"\nI0718 20:02:03.361318       1 event.go:291] \"Event occurred\" object=\"provisioning-7043-7873/csi-hostpath-resizer\" kind=\"StatefulSet\" apiVersion=\"apps/v1\" type=\"Normal\" reason=\"SuccessfulCreate\" message=\"create Pod csi-hostpath-resizer-0 in StatefulSet csi-hostpath-resizer successful\"\nI0718 20:02:03.452415       1 event.go:291] \"Event occurred\" object=\"provisioning-7043-7873/csi-hostpath-snapshotter\" kind=\"StatefulSet\" apiVersion=\"apps/v1\" type=\"Normal\" reason=\"SuccessfulCreate\" message=\"create Pod csi-hostpath-snapshotter-0 in StatefulSet csi-hostpath-snapshotter successful\"\nI0718 20:02:03.557078       1 event.go:291] \"Event occurred\" object=\"provisioning-7043/csi-hostpath7gbch\" kind=\"PersistentVolumeClaim\" apiVersion=\"v1\" type=\"Normal\" reason=\"ExternalProvisioning\" message=\"waiting for a volume to be created, either by external provisioner \\\"csi-hostpath-provisioning-7043\\\" or manually created by system administrator\"\nI0718 20:02:03.557137       1 event.go:291] \"Event occurred\" object=\"provisioning-7043/csi-hostpath7gbch\" kind=\"PersistentVolumeClaim\" apiVersion=\"v1\" type=\"Normal\" reason=\"ExternalProvisioning\" message=\"waiting for a volume to be created, either by external provisioner \\\"csi-hostpath-provisioning-7043\\\" or manually created by system administrator\"\nI0718 20:02:03.918905       1 namespace_controller.go:185] Namespace has been deleted subpath-4306\nI0718 20:02:04.329076       1 namespace_controller.go:185] Namespace has been deleted pv-protection-6080\nI0718 20:02:04.686013       1 namespace_controller.go:185] Namespace has been deleted dns-2451\nE0718 20:02:04.891473       1 reflector.go:138] k8s.io/client-go/metadata/metadatainformer/informer.go:90: Failed to watch *v1.PartialObjectMetadata: failed to list *v1.PartialObjectMetadata: the server could not find the requested resource\nE0718 20:02:04.932872       1 tokens_controller.go:262] error synchronizing serviceaccount volume-expand-5318-6478/default: secrets \"default-token-qfkjj\" is forbidden: unable to create new content in namespace volume-expand-5318-6478 because it is being terminated\nI0718 20:02:05.010354       1 namespace_controller.go:185] Namespace has been deleted csi-mock-volumes-1666\nI0718 20:02:05.253810       1 namespace_controller.go:185] Namespace has been deleted csi-mock-volumes-735-8114\nE0718 20:02:05.460313       1 reflector.go:138] k8s.io/client-go/metadata/metadatainformer/informer.go:90: Failed to watch *v1.PartialObjectMetadata: failed to list *v1.PartialObjectMetadata: the server could not find the requested resource\nI0718 20:02:05.533863       1 pv_controller.go:879] volume \"pvc-21f66de1-37b8-4329-a5b5-bd0271c6e945\" entered phase \"Bound\"\nI0718 20:02:05.533894       1 pv_controller.go:982] volume \"pvc-21f66de1-37b8-4329-a5b5-bd0271c6e945\" bound to claim \"provisioning-7043/csi-hostpath7gbch\"\nI0718 20:02:05.545971       1 pv_controller.go:823] claim \"provisioning-7043/csi-hostpath7gbch\" entered phase \"Bound\"\nE0718 20:02:05.601128       1 tokens_controller.go:262] error synchronizing serviceaccount kubectl-510/default: secrets \"default-token-zdzd8\" is forbidden: unable to create new content in namespace kubectl-510 because it is being terminated\nI0718 20:02:05.764459       1 reconciler.go:295] attacherDetacher.AttachVolume started for volume \"pvc-21f66de1-37b8-4329-a5b5-bd0271c6e945\" (UniqueName: \"kubernetes.io/csi/csi-hostpath-provisioning-7043^066c3db9-e803-11eb-8130-6692668a95ce\") from node \"10.116.0.7\" \nE0718 20:02:05.974399       1 tokens_controller.go:262] error synchronizing serviceaccount provisioning-4462/default: serviceaccounts \"default\" not found\nI0718 20:02:06.000969       1 deployment_controller.go:583] \"Deployment has been deleted\" deployment=\"webhook-2018/sample-webhook-deployment\"\nI0718 20:02:06.319931       1 operation_generator.go:368] AttachVolume.Attach succeeded for volume \"pvc-21f66de1-37b8-4329-a5b5-bd0271c6e945\" (UniqueName: \"kubernetes.io/csi/csi-hostpath-provisioning-7043^066c3db9-e803-11eb-8130-6692668a95ce\") from node \"10.116.0.7\" \nI0718 20:02:06.320071       1 event.go:291] \"Event occurred\" object=\"provisioning-7043/pod-subpath-test-dynamicpv-xdr6\" kind=\"Pod\" apiVersion=\"v1\" type=\"Normal\" reason=\"SuccessfulAttachVolume\" message=\"AttachVolume.Attach succeeded for volume \\\"pvc-21f66de1-37b8-4329-a5b5-bd0271c6e945\\\" \"\nI0718 20:02:06.568812       1 pv_controller.go:930] claim \"volume-6367/pvc-5kgs5\" bound to volume \"local-m8msb\"\nI0718 20:02:06.586880       1 pv_controller.go:879] volume \"local-m8msb\" entered phase \"Bound\"\nI0718 20:02:06.586908       1 pv_controller.go:982] volume \"local-m8msb\" bound to claim \"volume-6367/pvc-5kgs5\"\nI0718 20:02:06.599517       1 pv_controller.go:823] claim \"volume-6367/pvc-5kgs5\" entered phase \"Bound\"\nE0718 20:02:07.046984       1 tokens_controller.go:262] error synchronizing serviceaccount csi-mock-volumes-1666-4090/default: secrets \"default-token-mb7v6\" is forbidden: unable to create new content in namespace csi-mock-volumes-1666-4090 because it is being terminated\nE0718 20:02:07.529555       1 tokens_controller.go:262] error synchronizing serviceaccount cronjob-4272/default: serviceaccounts \"default\" not found\nI0718 20:02:07.644939       1 garbagecollector.go:471] \"Processing object\" object=\"provisioning-4462-6849/csi-hostpath-attacher-vpf8r\" objectUID=882e4796-01a6-4d47-a8dd-1ab722311cb7 kind=\"EndpointSlice\" virtual=false\nI0718 20:02:07.653239       1 garbagecollector.go:580] \"Deleting object\" object=\"provisioning-4462-6849/csi-hostpath-attacher-vpf8r\" objectUID=882e4796-01a6-4d47-a8dd-1ab722311cb7 kind=\"EndpointSlice\" propagationPolicy=Background\nI0718 20:02:07.696918       1 garbagecollector.go:471] \"Processing object\" object=\"provisioning-4462-6849/csi-hostpath-attacher-0\" objectUID=f9bb4ec4-d061-44d5-9a97-a41f0397de0c kind=\"Pod\" virtual=false\nI0718 20:02:07.696940       1 stateful_set.go:419] StatefulSet has been deleted provisioning-4462-6849/csi-hostpath-attacher\nI0718 20:02:07.697017       1 garbagecollector.go:471] \"Processing object\" object=\"provisioning-4462-6849/csi-hostpath-attacher-7f4f68d598\" objectUID=fa419ccf-f056-4e9f-beec-4b7cc35e4894 kind=\"ControllerRevision\" virtual=false\nI0718 20:02:07.699339       1 garbagecollector.go:580] \"Deleting object\" object=\"provisioning-4462-6849/csi-hostpath-attacher-0\" objectUID=f9bb4ec4-d061-44d5-9a97-a41f0397de0c kind=\"Pod\" propagationPolicy=Background\nI0718 20:02:07.700077       1 garbagecollector.go:580] \"Deleting object\" object=\"provisioning-4462-6849/csi-hostpath-attacher-7f4f68d598\" objectUID=fa419ccf-f056-4e9f-beec-4b7cc35e4894 kind=\"ControllerRevision\" propagationPolicy=Background\nI0718 20:02:07.771179       1 garbagecollector.go:471] \"Processing object\" object=\"provisioning-4462-6849/csi-hostpathplugin-nsrqt\" objectUID=ec40f802-da5c-4db8-bba5-b8d4a60c534f kind=\"EndpointSlice\" virtual=false\nI0718 20:02:07.775122       1 garbagecollector.go:580] \"Deleting object\" object=\"provisioning-4462-6849/csi-hostpathplugin-nsrqt\" objectUID=ec40f802-da5c-4db8-bba5-b8d4a60c534f kind=\"EndpointSlice\" propagationPolicy=Background\nI0718 20:02:07.816100       1 garbagecollector.go:471] \"Processing object\" object=\"provisioning-4462-6849/csi-hostpathplugin-6686c9c9fd\" objectUID=f0394b66-c3b2-4b34-9a36-5b2a4b236600 kind=\"ControllerRevision\" virtual=false\nI0718 20:02:07.816120       1 stateful_set.go:419] StatefulSet has been deleted provisioning-4462-6849/csi-hostpathplugin\nI0718 20:02:07.816173       1 garbagecollector.go:471] \"Processing object\" object=\"provisioning-4462-6849/csi-hostpathplugin-0\" objectUID=b01b03a6-9230-424e-9893-d263022258eb kind=\"Pod\" virtual=false\nI0718 20:02:07.820076       1 garbagecollector.go:580] \"Deleting object\" object=\"provisioning-4462-6849/csi-hostpathplugin-6686c9c9fd\" objectUID=f0394b66-c3b2-4b34-9a36-5b2a4b236600 kind=\"ControllerRevision\" propagationPolicy=Background\nI0718 20:02:07.820848       1 garbagecollector.go:580] \"Deleting object\" object=\"provisioning-4462-6849/csi-hostpathplugin-0\" objectUID=b01b03a6-9230-424e-9893-d263022258eb kind=\"Pod\" propagationPolicy=Background\nI0718 20:02:07.847518       1 garbagecollector.go:471] \"Processing object\" object=\"provisioning-4462-6849/csi-hostpath-provisioner-ndjnx\" objectUID=e69b32e5-8ede-4fc0-a5c3-8385599d9297 kind=\"EndpointSlice\" virtual=false\nI0718 20:02:07.849601       1 garbagecollector.go:580] \"Deleting object\" object=\"provisioning-4462-6849/csi-hostpath-provisioner-ndjnx\" objectUID=e69b32e5-8ede-4fc0-a5c3-8385599d9297 kind=\"EndpointSlice\" propagationPolicy=Background\nI0718 20:02:07.897633       1 stateful_set.go:419] StatefulSet has been deleted provisioning-4462-6849/csi-hostpath-provisioner\nI0718 20:02:07.897648       1 garbagecollector.go:471] \"Processing object\" object=\"provisioning-4462-6849/csi-hostpath-provisioner-c7d79d99b\" objectUID=8b8bb69b-a076-4e8b-9a37-1e919698de43 kind=\"ControllerRevision\" virtual=false\nI0718 20:02:07.897648       1 garbagecollector.go:471] \"Processing object\" object=\"provisioning-4462-6849/csi-hostpath-provisioner-0\" objectUID=928ebab0-5609-44c7-84d8-f972f2ed0b87 kind=\"Pod\" virtual=false\nI0718 20:02:07.899870       1 garbagecollector.go:580] \"Deleting object\" object=\"provisioning-4462-6849/csi-hostpath-provisioner-c7d79d99b\" objectUID=8b8bb69b-a076-4e8b-9a37-1e919698de43 kind=\"ControllerRevision\" propagationPolicy=Background\nI0718 20:02:07.900744       1 garbagecollector.go:580] \"Deleting object\" object=\"provisioning-4462-6849/csi-hostpath-provisioner-0\" objectUID=928ebab0-5609-44c7-84d8-f972f2ed0b87 kind=\"Pod\" propagationPolicy=Background\nI0718 20:02:07.931562       1 garbagecollector.go:471] \"Processing object\" object=\"provisioning-4462-6849/csi-hostpath-resizer-v2lcs\" objectUID=8aeffe08-c96e-47b6-8b06-6c6d6babe508 kind=\"EndpointSlice\" virtual=false\nI0718 20:02:07.935066       1 garbagecollector.go:580] \"Deleting object\" object=\"provisioning-4462-6849/csi-hostpath-resizer-v2lcs\" objectUID=8aeffe08-c96e-47b6-8b06-6c6d6babe508 kind=\"EndpointSlice\" propagationPolicy=Background\nI0718 20:02:07.984070       1 garbagecollector.go:471] \"Processing object\" object=\"provisioning-4462-6849/csi-hostpath-resizer-74f4f4bd55\" objectUID=3de1e9a2-78bb-4e82-83b5-f5e18acafa29 kind=\"ControllerRevision\" virtual=false\nI0718 20:02:07.984420       1 stateful_set.go:419] StatefulSet has been deleted provisioning-4462-6849/csi-hostpath-resizer\nI0718 20:02:07.984481       1 garbagecollector.go:471] \"Processing object\" object=\"provisioning-4462-6849/csi-hostpath-resizer-0\" objectUID=931a3857-4f17-4e57-8bac-52ba39db4530 kind=\"Pod\" virtual=false\nI0718 20:02:08.002767       1 garbagecollector.go:580] \"Deleting object\" object=\"provisioning-4462-6849/csi-hostpath-resizer-74f4f4bd55\" objectUID=3de1e9a2-78bb-4e82-83b5-f5e18acafa29 kind=\"ControllerRevision\" propagationPolicy=Background\nI0718 20:02:08.003404       1 garbagecollector.go:580] \"Deleting object\" object=\"provisioning-4462-6849/csi-hostpath-resizer-0\" objectUID=931a3857-4f17-4e57-8bac-52ba39db4530 kind=\"Pod\" propagationPolicy=Background\nI0718 20:02:08.014671       1 garbagecollector.go:471] \"Processing object\" object=\"provisioning-4462-6849/csi-hostpath-snapshotter-qxll4\" objectUID=f85c70d5-0f50-4d89-97f3-f9ccb77f75cf kind=\"EndpointSlice\" virtual=false\nI0718 20:02:08.033688       1 namespace_controller.go:185] Namespace has been deleted subpath-8417\nI0718 20:02:08.039910       1 garbagecollector.go:580] \"Deleting object\" object=\"provisioning-4462-6849/csi-hostpath-snapshotter-qxll4\" objectUID=f85c70d5-0f50-4d89-97f3-f9ccb77f75cf kind=\"EndpointSlice\" propagationPolicy=Background\nI0718 20:02:08.070331       1 garbagecollector.go:471] \"Processing object\" object=\"provisioning-4462-6849/csi-hostpath-snapshotter-84d8d49f79\" objectUID=66ced31f-e458-4165-be9d-168dd6ebe3fd kind=\"ControllerRevision\" virtual=false\nI0718 20:02:08.070361       1 garbagecollector.go:471] \"Processing object\" object=\"provisioning-4462-6849/csi-hostpath-snapshotter-0\" objectUID=42829f61-dfe8-47f1-97fd-504a3ef07544 kind=\"Pod\" virtual=false\nI0718 20:02:08.070464       1 stateful_set.go:419] StatefulSet has been deleted provisioning-4462-6849/csi-hostpath-snapshotter\nI0718 20:02:08.073656       1 garbagecollector.go:580] \"Deleting object\" object=\"provisioning-4462-6849/csi-hostpath-snapshotter-84d8d49f79\" objectUID=66ced31f-e458-4165-be9d-168dd6ebe3fd kind=\"ControllerRevision\" propagationPolicy=Background\nI0718 20:02:08.073797       1 garbagecollector.go:580] \"Deleting object\" object=\"provisioning-4462-6849/csi-hostpath-snapshotter-0\" objectUID=42829f61-dfe8-47f1-97fd-504a3ef07544 kind=\"Pod\" propagationPolicy=Background\nI0718 20:02:09.606941       1 namespace_controller.go:185] Namespace has been deleted pods-4456\nI0718 20:02:09.647863       1 pvc_protection_controller.go:291] \"PVC is unused\" PVC=\"provisioning-9310/pvc-49j46\"\nI0718 20:02:09.658539       1 pv_controller.go:640] volume \"local-5tr89\" is released and reclaim policy \"Retain\" will be executed\nI0718 20:02:09.667488       1 pv_controller.go:879] volume \"local-5tr89\" entered phase \"Released\"\nI0718 20:02:09.690953       1 pv_controller_base.go:505] deletion of claim \"provisioning-9310/pvc-49j46\" was already processed\nE0718 20:02:09.701941       1 reflector.go:138] k8s.io/client-go/metadata/metadatainformer/informer.go:90: Failed to watch *v1.PartialObjectMetadata: failed to list *v1.PartialObjectMetadata: the server could not find the requested resource\nE0718 20:02:09.792605       1 reflector.go:138] k8s.io/client-go/metadata/metadatainformer/informer.go:90: Failed to watch *v1.PartialObjectMetadata: failed to list *v1.PartialObjectMetadata: the server could not find the requested resource\nI0718 20:02:10.764864       1 namespace_controller.go:185] Namespace has been deleted kubectl-510\nI0718 20:02:10.804560       1 pvc_protection_controller.go:303] \"Pod uses PVC\" pod=\"persistent-local-volumes-test-7580/pod-18f8ee55-7e99-41d2-b8bd-6d08d446e074\" PVC=\"persistent-local-volumes-test-7580/pvc-mnzvt\"\nI0718 20:02:10.804583       1 pvc_protection_controller.go:181] \"Keeping PVC because it is being used\" PVC=\"persistent-local-volumes-test-7580/pvc-mnzvt\"\nI0718 20:02:10.928127       1 namespace_controller.go:185] Namespace has been deleted services-7615\nI0718 20:02:11.198383       1 namespace_controller.go:185] Namespace has been deleted provisioning-4462\nI0718 20:02:11.200083       1 replica_set.go:559] \"Too few replicas\" replicaSet=\"webhook-3379/sample-webhook-deployment-78988fc6cd\" need=1 creating=1\nI0718 20:02:11.200197       1 event.go:291] \"Event occurred\" object=\"webhook-3379/sample-webhook-deployment\" kind=\"Deployment\" apiVersion=\"apps/v1\" type=\"Normal\" reason=\"ScalingReplicaSet\" message=\"Scaled up replica set sample-webhook-deployment-78988fc6cd to 1\"\nI0718 20:02:11.210627       1 event.go:291] \"Event occurred\" object=\"webhook-3379/sample-webhook-deployment-78988fc6cd\" kind=\"ReplicaSet\" apiVersion=\"apps/v1\" type=\"Normal\" reason=\"SuccessfulCreate\" message=\"Created pod: sample-webhook-deployment-78988fc6cd-zmj2w\"\nI0718 20:02:11.212082       1 deployment_controller.go:490] \"Error syncing deployment\" deployment=\"webhook-3379/sample-webhook-deployment\" err=\"Operation cannot be fulfilled on deployments.apps \\\"sample-webhook-deployment\\\": the object has been modified; please apply your changes to the latest version and try again\"\nE0718 20:02:11.286479       1 tokens_controller.go:262] error synchronizing serviceaccount security-context-6721/default: secrets \"default-token-h2vrl\" is forbidden: unable to create new content in namespace security-context-6721 because it is being terminated\nI0718 20:02:11.615586       1 garbagecollector.go:471] \"Processing object\" object=\"webhook-7902/e2e-test-webhook-rb7pm\" objectUID=0b5d399d-b613-4d6c-a006-759430e616c4 kind=\"EndpointSlice\" virtual=false\nI0718 20:02:11.624107       1 garbagecollector.go:580] \"Deleting object\" object=\"webhook-7902/e2e-test-webhook-rb7pm\" objectUID=0b5d399d-b613-4d6c-a006-759430e616c4 kind=\"EndpointSlice\" propagationPolicy=Background\nI0718 20:02:11.685860       1 garbagecollector.go:471] \"Processing object\" object=\"webhook-7902/sample-webhook-deployment-78988fc6cd\" objectUID=60f8b77f-bbc5-41ac-8262-78523d8e3786 kind=\"ReplicaSet\" virtual=false\nI0718 20:02:11.686150       1 deployment_controller.go:583] \"Deployment has been deleted\" deployment=\"webhook-7902/sample-webhook-deployment\"\nI0718 20:02:11.688359       1 garbagecollector.go:580] \"Deleting object\" object=\"webhook-7902/sample-webhook-deployment-78988fc6cd\" objectUID=60f8b77f-bbc5-41ac-8262-78523d8e3786 kind=\"ReplicaSet\" propagationPolicy=Background\nI0718 20:02:11.697166       1 garbagecollector.go:471] \"Processing object\" object=\"webhook-7902/sample-webhook-deployment-78988fc6cd-d4hgx\" objectUID=e6a2f6cf-81a1-4a68-9396-241c6d7bc420 kind=\"Pod\" virtual=false\nI0718 20:02:11.699835       1 garbagecollector.go:580] \"Deleting object\" object=\"webhook-7902/sample-webhook-deployment-78988fc6cd-d4hgx\" objectUID=e6a2f6cf-81a1-4a68-9396-241c6d7bc420 kind=\"Pod\" propagationPolicy=Background\nI0718 20:02:12.725803       1 pvc_protection_controller.go:291] \"PVC is unused\" PVC=\"volume-7899/pvc-kxxc4\"\nI0718 20:02:12.736738       1 pv_controller.go:640] volume \"local-vrxvf\" is released and reclaim policy \"Retain\" will be executed\nI0718 20:02:12.741077       1 pv_controller.go:879] volume \"local-vrxvf\" entered phase \"Released\"\nI0718 20:02:12.765962       1 pv_controller_base.go:505] deletion of claim \"volume-7899/pvc-kxxc4\" was already processed\nE0718 20:02:12.785146       1 tokens_controller.go:262] error synchronizing serviceaccount pod-disks-3683/default: secrets \"default-token-kf2mt\" is forbidden: unable to create new content in namespace pod-disks-3683 because it is being terminated\nE0718 20:02:13.218967       1 tokens_controller.go:262] error synchronizing serviceaccount replicaset-5128/default: secrets \"default-token-pxx7r\" is forbidden: unable to create new content in namespace replicaset-5128 because it is being terminated\nI0718 20:02:13.227691       1 namespace_controller.go:185] Namespace has been deleted services-8331\nE0718 20:02:14.156160       1 reflector.go:138] k8s.io/client-go/metadata/metadatainformer/informer.go:90: Failed to watch *v1.PartialObjectMetadata: failed to list *v1.PartialObjectMetadata: the server could not find the requested resource\nI0718 20:02:16.142052       1 namespace_controller.go:185] Namespace has been deleted kubectl-1946\nI0718 20:02:16.397668       1 namespace_controller.go:185] Namespace has been deleted security-context-6721\nE0718 20:02:16.443855       1 reflector.go:138] k8s.io/client-go/metadata/metadatainformer/informer.go:90: Failed to watch *v1.PartialObjectMetadata: failed to list *v1.PartialObjectMetadata: the server could not find the requested resource\nE0718 20:02:16.582665       1 tokens_controller.go:262] error synchronizing serviceaccount webhook-7902/default: secrets \"default-token-cb8bq\" is forbidden: unable to create new content in namespace webhook-7902 because it is being terminated\nE0718 20:02:16.776822       1 tokens_controller.go:262] error synchronizing serviceaccount persistent-local-volumes-test-7580/default: secrets \"default-token-l25pw\" is forbidden: unable to create new content in namespace persistent-local-volumes-test-7580 because it is being terminated\nI0718 20:02:16.829679       1 pvc_protection_controller.go:291] \"PVC is unused\" PVC=\"persistent-local-volumes-test-7580/pvc-mnzvt\"\nI0718 20:02:16.844153       1 pv_controller.go:640] volume \"local-pvf4rsz\" is released and reclaim policy \"Retain\" will be executed\nI0718 20:02:16.852286       1 pv_controller.go:879] volume \"local-pvf4rsz\" entered phase \"Released\"\nI0718 20:02:16.859306       1 pv_controller_base.go:505] deletion of claim \"persistent-local-volumes-test-7580/pvc-mnzvt\" was already processed\nI0718 20:02:17.620756       1 namespace_controller.go:185] Namespace has been deleted kubectl-9000\nI0718 20:02:17.893002       1 namespace_controller.go:185] Namespace has been deleted pod-disks-3683\nI0718 20:02:18.345621       1 namespace_controller.go:185] Namespace has been deleted replicaset-5128\nE0718 20:02:18.365147       1 namespace_controller.go:162] deletion of namespace kubectl-9687 failed: unable to retrieve the complete list of server APIs: crd-publish-openapi-test-common-group.example.com/v6: the server could not find the requested resource\nE0718 20:02:18.542317       1 namespace_controller.go:162] deletion of namespace kubectl-9687 failed: unable to retrieve the complete list of server APIs: crd-publish-openapi-test-common-group.example.com/v6: the server could not find the requested resource\nI0718 20:02:18.622838       1 namespace_controller.go:185] Namespace has been deleted projected-4120\nE0718 20:02:18.716292       1 namespace_controller.go:162] deletion of namespace kubectl-9687 failed: unable to retrieve the complete list of server APIs: crd-publish-openapi-test-common-group.example.com/v6: the server could not find the requested resource\nE0718 20:02:18.757564       1 namespace_controller.go:162] deletion of namespace provisioning-4462-6849 failed: unable to retrieve the complete list of server APIs: crd-publish-openapi-test-common-group.example.com/v6: the server could not find the requested resource\nE0718 20:02:18.822145       1 tokens_controller.go:262] error synchronizing serviceaccount volume-7899/default: secrets \"default-token-7w9k2\" is forbidden: unable to create new content in namespace volume-7899 because it is being terminated\nI0718 20:02:18.825794       1 utils.go:366] couldn't find ipfamilies for headless service: services-4562/clusterip-service likely because controller manager is likely connected to an old apiserver that does not support ip families yet. The service endpoint slice will use dual stack families until api-server default it correctly\nE0718 20:02:18.936367       1 namespace_controller.go:162] deletion of namespace volume-7899 failed: unable to retrieve the complete list of server APIs: crd-publish-openapi-test-common-group.example.com/v6: the server could not find the requested resource\nE0718 20:02:18.938630       1 namespace_controller.go:162] deletion of namespace kubectl-9687 failed: unable to retrieve the complete list of server APIs: crd-publish-openapi-test-common-group.example.com/v6: the server could not find the requested resource\nE0718 20:02:18.966649       1 namespace_controller.go:162] deletion of namespace provisioning-4462-6849 failed: unable to retrieve the complete list of server APIs: crd-publish-openapi-test-common-group.example.com/v6: the server could not find the requested resource\nE0718 20:02:19.111846       1 namespace_controller.go:162] deletion of namespace volume-7899 failed: unable to retrieve the complete list of server APIs: crd-publish-openapi-test-common-group.example.com/v6: the server could not find the requested resource\nE0718 20:02:19.200167       1 namespace_controller.go:162] deletion of namespace kubectl-9687 failed: unable to retrieve the complete list of server APIs: crd-publish-openapi-test-common-group.example.com/v6: the server could not find the requested resource\nE0718 20:02:19.208300       1 namespace_controller.go:162] deletion of namespace provisioning-4462-6849 failed: unable to retrieve the complete list of server APIs: crd-publish-openapi-test-common-group.example.com/v6: the server could not find the requested resource\nE0718 20:02:19.312560       1 namespace_controller.go:162] deletion of namespace volume-7899 failed: unable to retrieve the complete list of server APIs: crd-publish-openapi-test-common-group.example.com/v6: the server could not find the requested resource\nE0718 20:02:19.401582       1 namespace_controller.go:162] deletion of namespace provisioning-4462-6849 failed: unable to retrieve the complete list of server APIs: crd-publish-openapi-test-common-group.example.com/v6: the server could not find the requested resource\nE0718 20:02:19.457915       1 namespace_controller.go:162] deletion of namespace kubectl-9687 failed: unable to retrieve the complete list of server APIs: crd-publish-openapi-test-common-group.example.com/v6: the server could not find the requested resource\nE0718 20:02:19.515443       1 namespace_controller.go:162] deletion of namespace volume-7899 failed: unable to retrieve the complete list of server APIs: crd-publish-openapi-test-common-group.example.com/v6: the server could not find the requested resource\nE0718 20:02:19.603383       1 namespace_controller.go:162] deletion of namespace provisioning-4462-6849 failed: unable to retrieve the complete list of server APIs: crd-publish-openapi-test-common-group.example.com/v6: the server could not find the requested resource\nE0718 20:02:19.702699       1 namespace_controller.go:162] deletion of namespace volume-7899 failed: unable to retrieve the complete list of server APIs: crd-publish-openapi-test-common-group.example.com/v6: the server could not find the requested resource\nE0718 20:02:19.789548       1 namespace_controller.go:162] deletion of namespace kubectl-9687 failed: unable to retrieve the complete list of server APIs: crd-publish-openapi-test-common-group.example.com/v6: the server could not find the requested resource\nI0718 20:02:19.832705       1 utils.go:366] couldn't find ipfamilies for headless service: services-4562/clusterip-service likely because controller manager is likely connected to an old apiserver that does not support ip families yet. The service endpoint slice will use dual stack families until api-server default it correctly\nE0718 20:02:19.858701       1 namespace_controller.go:162] deletion of namespace provisioning-4462-6849 failed: unable to retrieve the complete list of server APIs: crd-publish-openapi-test-common-group.example.com/v6: the server could not find the requested resource\nE0718 20:02:19.928609       1 namespace_controller.go:162] deletion of namespace volume-7899 failed: unable to retrieve the complete list of server APIs: crd-publish-openapi-test-common-group.example.com/v6: the server could not find the requested resource\nI0718 20:02:20.132497       1 pvc_protection_controller.go:303] \"Pod uses PVC\" pod=\"ephemeral-5827/inline-volume-tester2-qw2vr\" PVC=\"ephemeral-5827/inline-volume-tester2-qw2vr-my-volume-0\"\nI0718 20:02:20.132521       1 pvc_protection_controller.go:181] \"Keeping PVC because it is being used\" PVC=\"ephemeral-5827/inline-volume-tester2-qw2vr-my-volume-0\"\nI0718 20:02:20.173147       1 reconciler.go:219] attacherDetacher.DetachVolume started for volume \"pvc-cd083b17-8ca9-40cf-80e6-aefca09f6ef7\" (UniqueName: \"kubernetes.io/csi/csi-hostpath-ephemeral-5827^da7cbc4f-e802-11eb-8233-02fc2257e70c\") on node \"10.116.0.7\" \nI0718 20:02:20.180515       1 operation_generator.go:1483] Verified volume is safe to detach for volume \"pvc-cd083b17-8ca9-40cf-80e6-aefca09f6ef7\" (UniqueName: \"kubernetes.io/csi/csi-hostpath-ephemeral-5827^da7cbc4f-e802-11eb-8233-02fc2257e70c\") on node \"10.116.0.7\" \nE0718 20:02:20.196678       1 namespace_controller.go:162] deletion of namespace provisioning-4462-6849 failed: unable to retrieve the complete list of server APIs: crd-publish-openapi-test-common-group.example.com/v6: the server could not find the requested resource\nE0718 20:02:20.256080       1 namespace_controller.go:162] deletion of namespace volume-7899 failed: unable to retrieve the complete list of server APIs: crd-publish-openapi-test-common-group.example.com/v6: the server could not find the requested resource\nE0718 20:02:20.304173       1 namespace_controller.go:162] deletion of namespace kubectl-9687 failed: unable to retrieve the complete list of server APIs: crd-publish-openapi-test-common-group.example.com/v6: the server could not find the requested resource\nI0718 20:02:20.336947       1 pvc_protection_controller.go:291] \"PVC is unused\" PVC=\"ephemeral-5827/inline-volume-tester2-qw2vr-my-volume-0\"\nI0718 20:02:20.346710       1 namespace_controller.go:185] Namespace has been deleted provisioning-9310\nI0718 20:02:20.346761       1 garbagecollector.go:471] \"Processing object\" object=\"ephemeral-5827/inline-volume-tester2-qw2vr\" objectUID=fee9b4b0-5efd-4b4e-8164-bfc700bc2c63 kind=\"Pod\" virtual=false\nI0718 20:02:20.349416       1 garbagecollector.go:590] remove DeleteDependents finalizer for item [v1/Pod, namespace: ephemeral-5827, name: inline-volume-tester2-qw2vr, uid: fee9b4b0-5efd-4b4e-8164-bfc700bc2c63]\nI0718 20:02:20.350030       1 pv_controller.go:640] volume \"pvc-cd083b17-8ca9-40cf-80e6-aefca09f6ef7\" is released and reclaim policy \"Delete\" will be executed\nI0718 20:02:20.356008       1 pv_controller.go:879] volume \"pvc-cd083b17-8ca9-40cf-80e6-aefca09f6ef7\" entered phase \"Released\"\nI0718 20:02:20.358505       1 pv_controller.go:1341] isVolumeReleased[pvc-cd083b17-8ca9-40cf-80e6-aefca09f6ef7]: volume is released\nI0718 20:02:20.386224       1 pv_controller_base.go:505] deletion of claim \"ephemeral-5827/inline-volume-tester2-qw2vr-my-volume-0\" was already processed\nI0718 20:02:20.616840       1 replica_set.go:559] \"Too few replicas\" replicaSet=\"services-2384/affinity-clusterip-timeout\" need=3 creating=3\nI0718 20:02:20.627360       1 event.go:291] \"Event occurred\" object=\"services-2384/affinity-clusterip-timeout\" kind=\"ReplicationController\" apiVersion=\"v1\" type=\"Normal\" reason=\"SuccessfulCreate\" message=\"Created pod: affinity-clusterip-timeout-62v7b\"\nI0718 20:02:20.640636       1 event.go:291] \"Event occurred\" object=\"services-2384/affinity-clusterip-timeout\" kind=\"ReplicationController\" apiVersion=\"v1\" type=\"Normal\" reason=\"SuccessfulCreate\" message=\"Created pod: affinity-clusterip-timeout-xjbkw\"\nI0718 20:02:20.643654       1 event.go:291] \"Event occurred\" object=\"services-2384/affinity-clusterip-timeout\" kind=\"ReplicationController\" apiVersion=\"v1\" type=\"Normal\" reason=\"SuccessfulCreate\" message=\"Created pod: affinity-clusterip-timeout-gm46x\"\nE0718 20:02:20.720850       1 namespace_controller.go:162] deletion of namespace provisioning-1806-9608 failed: unable to retrieve the complete list of server APIs: crd-publish-openapi-test-common-group.example.com/v6: the server could not find the requested resource\nI0718 20:02:20.729993       1 operation_generator.go:483] DetachVolume.Detach succeeded for volume \"pvc-cd083b17-8ca9-40cf-80e6-aefca09f6ef7\" (UniqueName: \"kubernetes.io/csi/csi-hostpath-ephemeral-5827^da7cbc4f-e802-11eb-8233-02fc2257e70c\") on node \"10.116.0.7\" \nE0718 20:02:20.745291       1 namespace_controller.go:162] deletion of namespace provisioning-4462-6849 failed: unable to retrieve the complete list of server APIs: crd-publish-openapi-test-common-group.example.com/v6: the server could not find the requested resource\nE0718 20:02:20.791752       1 namespace_controller.go:162] deletion of namespace volume-7899 failed: unable to retrieve the complete list of server APIs: crd-publish-openapi-test-common-group.example.com/v6: the server could not find the requested resource\nE0718 20:02:20.878297       1 namespace_controller.go:162] deletion of namespace provisioning-1806-9608 failed: unable to retrieve the complete list of server APIs: crd-publish-openapi-test-common-group.example.com/v6: the server could not find the requested resource\nI0718 20:02:20.945530       1 namespace_controller.go:185] Namespace has been deleted volume-expand-5318-6478\nI0718 20:02:20.991501       1 garbagecollector.go:471] \"Processing object\" object=\"webhook-3379/e2e-test-webhook-9k6l4\" objectUID=969de535-ccd3-49f3-a7fe-f5d8ada49b3e kind=\"EndpointSlice\" virtual=false\nI0718 20:02:20.996159       1 garbagecollector.go:580] \"Deleting object\" object=\"webhook-3379/e2e-test-webhook-9k6l4\" objectUID=969de535-ccd3-49f3-a7fe-f5d8ada49b3e kind=\"EndpointSlice\" propagationPolicy=Background\nI0718 20:02:21.031998       1 pv_controller.go:879] volume \"local-pvxxhrq\" entered phase \"Available\"\nI0718 20:02:21.036565       1 garbagecollector.go:471] \"Processing object\" object=\"webhook-3379/sample-webhook-deployment-78988fc6cd\" objectUID=07c2eb64-1f92-41e0-838f-cf3956ab8cc0 kind=\"ReplicaSet\" virtual=false\nI0718 20:02:21.036600       1 deployment_controller.go:583] \"Deployment has been deleted\" deployment=\"webhook-3379/sample-webhook-deployment\"\nI0718 20:02:21.039161       1 garbagecollector.go:580] \"Deleting object\" object=\"webhook-3379/sample-webhook-deployment-78988fc6cd\" objectUID=07c2eb64-1f92-41e0-838f-cf3956ab8cc0 kind=\"ReplicaSet\" propagationPolicy=Background\nI0718 20:02:21.048019       1 garbagecollector.go:471] \"Processing object\" object=\"webhook-3379/sample-webhook-deployment-78988fc6cd-zmj2w\" objectUID=7cd0f4b1-7213-4423-8e4c-13a07585194b kind=\"Pod\" virtual=false\nI0718 20:02:21.050307       1 garbagecollector.go:580] \"Deleting object\" object=\"webhook-3379/sample-webhook-deployment-78988fc6cd-zmj2w\" objectUID=7cd0f4b1-7213-4423-8e4c-13a07585194b kind=\"Pod\" propagationPolicy=Background\nI0718 20:02:21.059650       1 pv_controller.go:930] claim \"persistent-local-volumes-test-448/pvc-4frjz\" bound to volume \"local-pvxxhrq\"\nI0718 20:02:21.071446       1 pv_controller.go:879] volume \"local-pvxxhrq\" entered phase \"Bound\"\nI0718 20:02:21.071478       1 pv_controller.go:982] volume \"local-pvxxhrq\" bound to claim \"persistent-local-volumes-test-448/pvc-4frjz\"\nI0718 20:02:21.084024       1 pv_controller.go:823] claim \"persistent-local-volumes-test-448/pvc-4frjz\" entered phase \"Bound\"\nE0718 20:02:21.094394       1 namespace_controller.go:162] deletion of namespace provisioning-1806-9608 failed: unable to retrieve the complete list of server APIs: crd-publish-openapi-test-common-group.example.com/v6: the server could not find the requested resource\nE0718 20:02:21.111640       1 reflector.go:138] k8s.io/client-go/metadata/metadatainformer/informer.go:90: Failed to watch *v1.PartialObjectMetadata: failed to list *v1.PartialObjectMetadata: the server could not find the requested resource\nE0718 20:02:21.164153       1 namespace_controller.go:162] deletion of namespace kubectl-9687 failed: unable to retrieve the complete list of server APIs: crd-publish-openapi-test-common-group.example.com/v6: the server could not find the requested resource\nI0718 20:02:21.226817       1 pvc_protection_controller.go:291] \"PVC is unused\" PVC=\"persistent-local-volumes-test-448/pvc-4frjz\"\nI0718 20:02:21.235521       1 pv_controller.go:640] volume \"local-pvxxhrq\" is released and reclaim policy \"Retain\" will be executed\nI0718 20:02:21.240882       1 pv_controller.go:879] volume \"local-pvxxhrq\" entered phase \"Released\"\nI0718 20:02:21.269430       1 pv_controller_base.go:505] deletion of claim \"persistent-local-volumes-test-448/pvc-4frjz\" was already processed\nE0718 20:02:21.304019       1 namespace_controller.go:162] deletion of namespace provisioning-1806-9608 failed: unable to retrieve the complete list of server APIs: crd-publish-openapi-test-common-group.example.com/v6: the server could not find the requested resource\nE0718 20:02:21.522975       1 namespace_controller.go:162] deletion of namespace provisioning-1806-9608 failed: unable to retrieve the complete list of server APIs: crd-publish-openapi-test-common-group.example.com/v6: the server could not find the requested resource\nE0718 20:02:21.565631       1 namespace_controller.go:162] deletion of namespace provisioning-4462-6849 failed: unable to retrieve the complete list of server APIs: crd-publish-openapi-test-common-group.example.com/v6: the server could not find the requested resource\nE0718 20:02:21.586003       1 namespace_controller.go:162] deletion of namespace volume-7899 failed: unable to retrieve the complete list of server APIs: crd-publish-openapi-test-common-group.example.com/v6: the server could not find the requested resource\nI0718 20:02:21.647656       1 graph_builder.go:587] add [v1/Pod, namespace: ephemeral-5827, name: inline-volume-tester-knktf, uid: 9aebdd3c-8ab2-4b18-9157-091e2417cf4a] to the attemptToDelete, because it's waiting for its dependents to be deleted\nI0718 20:02:21.647707       1 garbagecollector.go:471] \"Processing object\" object=\"ephemeral-5827/inline-volume-tester-knktf-my-volume-0\" objectUID=dc473523-d38a-4923-8439-32b2189b8860 kind=\"PersistentVolumeClaim\" virtual=false\nI0718 20:02:21.647774       1 garbagecollector.go:471] \"Processing object\" object=\"ephemeral-5827/inline-volume-tester-knktf\" objectUID=9aebdd3c-8ab2-4b18-9157-091e2417cf4a kind=\"Pod\" virtual=false\nI0718 20:02:21.653341       1 garbagecollector.go:595] adding [v1/PersistentVolumeClaim, namespace: ephemeral-5827, name: inline-volume-tester-knktf-my-volume-0, uid: dc473523-d38a-4923-8439-32b2189b8860] to attemptToDelete, because its owner [v1/Pod, namespace: ephemeral-5827, name: inline-volume-tester-knktf, uid: 9aebdd3c-8ab2-4b18-9157-091e2417cf4a] is deletingDependents\nI0718 20:02:21.657312       1 garbagecollector.go:580] \"Deleting object\" object=\"ephemeral-5827/inline-volume-tester-knktf-my-volume-0\" objectUID=dc473523-d38a-4923-8439-32b2189b8860 kind=\"PersistentVolumeClaim\" propagationPolicy=Background\nI0718 20:02:21.668120       1 garbagecollector.go:471] \"Processing object\" object=\"ephemeral-5827/inline-volume-tester-knktf-my-volume-0\" objectUID=dc473523-d38a-4923-8439-32b2189b8860 kind=\"PersistentVolumeClaim\" virtual=false\nI0718 20:02:21.668145       1 pvc_protection_controller.go:303] \"Pod uses PVC\" pod=\"ephemeral-5827/inline-volume-tester-knktf\" PVC=\"ephemeral-5827/inline-volume-tester-knktf-my-volume-0\"\nI0718 20:02:21.668162       1 pvc_protection_controller.go:181] \"Keeping PVC because it is being used\" PVC=\"ephemeral-5827/inline-volume-tester-knktf-my-volume-0\"\nI0718 20:02:21.787400       1 namespace_controller.go:185] Namespace has been deleted webhook-7902-markers\nE0718 20:02:21.805873       1 namespace_controller.go:162] deletion of namespace provisioning-1806-9608 failed: unable to retrieve the complete list of server APIs: crd-publish-openapi-test-common-group.example.com/v6: the server could not find the requested resource\nI0718 20:02:21.809040       1 namespace_controller.go:185] Namespace has been deleted webhook-7902\nI0718 20:02:22.012029       1 graph_builder.go:587] add [v1/Pod, namespace: gc-7976, name: pod1, uid: 758b4901-1ae6-48f8-b098-e8e42d8d93d7] to the attemptToDelete, because it's waiting for its dependents to be deleted\nI0718 20:02:22.012090       1 garbagecollector.go:471] \"Processing object\" object=\"gc-7976/pod2\" objectUID=4f9f5773-a978-4539-a511-555af10cadf6 kind=\"Pod\" virtual=false\nI0718 20:02:22.012139       1 garbagecollector.go:471] \"Processing object\" object=\"gc-7976/pod1\" objectUID=758b4901-1ae6-48f8-b098-e8e42d8d93d7 kind=\"Pod\" virtual=false\nI0718 20:02:22.015221       1 garbagecollector.go:595] adding [v1/Pod, namespace: gc-7976, name: pod2, uid: 4f9f5773-a978-4539-a511-555af10cadf6] to attemptToDelete, because its owner [v1/Pod, namespace: gc-7976, name: pod1, uid: 758b4901-1ae6-48f8-b098-e8e42d8d93d7] is deletingDependents\nI0718 20:02:22.017619       1 garbagecollector.go:556] at least one owner of object [v1/Pod, namespace: gc-7976, name: pod2, uid: 4f9f5773-a978-4539-a511-555af10cadf6] has FinalizerDeletingDependents, and the object itself has dependents, so it is going to be deleted in Foreground\nI0718 20:02:22.025157       1 garbagecollector.go:471] \"Processing object\" object=\"gc-7976/pod2\" objectUID=4f9f5773-a978-4539-a511-555af10cadf6 kind=\"Pod\" virtual=false\nI0718 20:02:22.025198       1 graph_builder.go:587] add [v1/Pod, namespace: gc-7976, name: pod2, uid: 4f9f5773-a978-4539-a511-555af10cadf6] to the attemptToDelete, because it's waiting for its dependents to be deleted\nI0718 20:02:22.025241       1 garbagecollector.go:471] \"Processing object\" object=\"gc-7976/pod3\" objectUID=981b152f-89f4-4e3c-9acd-271a2c9e6015 kind=\"Pod\" virtual=false\nI0718 20:02:22.027492       1 garbagecollector.go:595] adding [v1/Pod, namespace: gc-7976, name: pod3, uid: 981b152f-89f4-4e3c-9acd-271a2c9e6015] to attemptToDelete, because its owner [v1/Pod, namespace: gc-7976, name: pod2, uid: 4f9f5773-a978-4539-a511-555af10cadf6] is deletingDependents\nI0718 20:02:22.027535       1 garbagecollector.go:471] \"Processing object\" object=\"gc-7976/pod2\" objectUID=4f9f5773-a978-4539-a511-555af10cadf6 kind=\"Pod\" virtual=false\nI0718 20:02:22.029894       1 garbagecollector.go:595] adding [v1/Pod, namespace: gc-7976, name: pod3, uid: 981b152f-89f4-4e3c-9acd-271a2c9e6015] to attemptToDelete, because its owner [v1/Pod, namespace: gc-7976, name: pod2, uid: 4f9f5773-a978-4539-a511-555af10cadf6] is deletingDependents\nI0718 20:02:22.030754       1 garbagecollector.go:545] processing object [v1/Pod, namespace: gc-7976, name: pod3, uid: 981b152f-89f4-4e3c-9acd-271a2c9e6015], some of its owners and its dependent [[v1/Pod, namespace: gc-7976, name: pod1, uid: 758b4901-1ae6-48f8-b098-e8e42d8d93d7]] have FinalizerDeletingDependents, to prevent potential cycle, its ownerReferences are going to be modified to be non-blocking, then the object is going to be deleted with Foreground\nI0718 20:02:22.036370       1 garbagecollector.go:556] at least one owner of object [v1/Pod, namespace: gc-7976, name: pod3, uid: 981b152f-89f4-4e3c-9acd-271a2c9e6015] has FinalizerDeletingDependents, and the object itself has dependents, so it is going to be deleted in Foreground\nI0718 20:02:22.036510       1 garbagecollector.go:471] \"Processing object\" object=\"gc-7976/pod2\" objectUID=4f9f5773-a978-4539-a511-555af10cadf6 kind=\"Pod\" virtual=false\nI0718 20:02:22.039343       1 garbagecollector.go:590] remove DeleteDependents finalizer for item [v1/Pod, namespace: gc-7976, name: pod2, uid: 4f9f5773-a978-4539-a511-555af10cadf6]\nI0718 20:02:22.041247       1 namespace_controller.go:185] Namespace has been deleted persistent-local-volumes-test-7580\nI0718 20:02:22.042860       1 garbagecollector.go:471] \"Processing object\" object=\"gc-7976/pod3\" objectUID=981b152f-89f4-4e3c-9acd-271a2c9e6015 kind=\"Pod\" virtual=false\nI0718 20:02:22.042886       1 graph_builder.go:587] add [v1/Pod, namespace: gc-7976, name: pod3, uid: 981b152f-89f4-4e3c-9acd-271a2c9e6015] to the attemptToDelete, because it's waiting for its dependents to be deleted\nI0718 20:02:22.042924       1 garbagecollector.go:471] \"Processing object\" object=\"gc-7976/pod1\" objectUID=758b4901-1ae6-48f8-b098-e8e42d8d93d7 kind=\"Pod\" virtual=false\nI0718 20:02:22.046308       1 garbagecollector.go:471] \"Processing object\" object=\"gc-7976/pod3\" objectUID=981b152f-89f4-4e3c-9acd-271a2c9e6015 kind=\"Pod\" virtual=false\nI0718 20:02:22.056347       1 garbagecollector.go:471] \"Processing object\" object=\"gc-7976/pod3\" objectUID=981b152f-89f4-4e3c-9acd-271a2c9e6015 kind=\"Pod\" virtual=false\nI0718 20:02:22.056381       1 garbagecollector.go:471] \"Processing object\" object=\"gc-7976/pod1\" objectUID=758b4901-1ae6-48f8-b098-e8e42d8d93d7 kind=\"Pod\" virtual=false\nI0718 20:02:22.059419       1 garbagecollector.go:590] remove DeleteDependents finalizer for item [v1/Pod, namespace: gc-7976, name: pod1, uid: 758b4901-1ae6-48f8-b098-e8e42d8d93d7]\nI0718 20:02:22.070871       1 garbagecollector.go:471] \"Processing object\" object=\"gc-7976/pod3\" objectUID=981b152f-89f4-4e3c-9acd-271a2c9e6015 kind=\"Pod\" virtual=false\nI0718 20:02:22.073144       1 garbagecollector.go:590] remove DeleteDependents finalizer for item [v1/Pod, namespace: gc-7976, name: pod3, uid: 981b152f-89f4-4e3c-9acd-271a2c9e6015]\nE0718 20:02:22.158358       1 namespace_controller.go:162] deletion of namespace provisioning-1806-9608 failed: unable to retrieve the complete list of server APIs: crd-publish-openapi-test-common-group.example.com/v6: the server could not find the requested resource\nI0718 20:02:22.177518       1 namespace_controller.go:185] Namespace has been deleted csi-mock-volumes-5970-9198\nE0718 20:02:22.366781       1 resource_quota_controller.go:409] unable to retrieve the complete list of server APIs: crd-publish-openapi-test-common-group.example.com/v6: the server could not find the requested resource\nE0718 20:02:22.592667       1 namespace_controller.go:162] deletion of namespace kubectl-9687 failed: unable to retrieve the complete list of server APIs: crd-publish-openapi-test-common-group.example.com/v6: the server could not find the requested resource\nE0718 20:02:22.619305       1 namespace_controller.go:162] deletion of namespace provisioning-1806-9608 failed: unable to retrieve the complete list of server APIs: crd-publish-openapi-test-common-group.example.com/v6: the server could not find the requested resource\nI0718 20:02:22.808426       1 namespace_controller.go:185] Namespace has been deleted csi-mock-volumes-1666-4090\nW0718 20:02:22.843825       1 garbagecollector.go:703] failed to discover some groups: map[crd-publish-openapi-test-common-group.example.com/v6:the server could not find the requested resource]\nI0718 20:02:22.843935       1 garbagecollector.go:213] syncing garbage collector with updated resources from discovery (attempt 1): added: [], removed: [kubectl.example.com/v1, Resource=e2e-test-kubectl-4850-crds]\nI0718 20:02:22.843992       1 shared_informer.go:240] Waiting for caches to sync for garbage collector\nI0718 20:02:22.844030       1 shared_informer.go:247] Caches are synced for garbage collector \nI0718 20:02:22.844037       1 garbagecollector.go:254] synced garbage collector\nE0718 20:02:23.010104       1 namespace_controller.go:162] deletion of namespace provisioning-4462-6849 failed: unable to retrieve the complete list of server APIs: crd-publish-openapi-test-common-group.example.com/v6: the server could not find the requested resource\nE0718 20:02:23.014335       1 namespace_controller.go:162] deletion of namespace volume-7899 failed: unable to retrieve the complete list of server APIs: crd-publish-openapi-test-common-group.example.com/v6: the server could not find the requested resource\nE0718 20:02:23.206695       1 reflector.go:138] k8s.io/client-go/metadata/metadatainformer/informer.go:90: Failed to watch *v1.PartialObjectMetadata: failed to list *v1.PartialObjectMetadata: the server could not find the requested resource\nE0718 20:02:24.965105       1 reflector.go:138] k8s.io/client-go/metadata/metadatainformer/informer.go:90: Failed to watch *v1.PartialObjectMetadata: failed to list *v1.PartialObjectMetadata: the server could not find the requested resource\nI0718 20:02:25.000475       1 deployment_controller.go:583] \"Deployment has been deleted\" deployment=\"webhook-4829/sample-webhook-deployment\"\nE0718 20:02:26.118307       1 tokens_controller.go:262] error synchronizing serviceaccount emptydir-7240/default: secrets \"default-token-lbj28\" is forbidden: unable to create new content in namespace emptydir-7240 because it is being terminated\nE0718 20:02:27.103122       1 tokens_controller.go:262] error synchronizing serviceaccount persistent-local-volumes-test-448/default: secrets \"default-token-nqk9c\" is forbidden: unable to create new content in namespace persistent-local-volumes-test-448 because it is being terminated\nE0718 20:02:27.988426       1 reflector.go:138] k8s.io/client-go/metadata/metadatainformer/informer.go:90: Failed to watch *v1.PartialObjectMetadata: failed to list *v1.PartialObjectMetadata: the server could not find the requested resource\nI0718 20:02:28.061792       1 pv_controller.go:879] volume \"local-pvkl66h\" entered phase \"Available\"\nI0718 20:02:28.088421       1 pv_controller.go:930] claim \"persistent-local-volumes-test-8575/pvc-p4ps5\" bound to volume \"local-pvkl66h\"\nI0718 20:02:28.099099       1 pv_controller.go:879] volume \"local-pvkl66h\" entered phase \"Bound\"\nI0718 20:02:28.099119       1 pv_controller.go:982] volume \"local-pvkl66h\" bound to claim \"persistent-local-volumes-test-8575/pvc-p4ps5\"\nI0718 20:02:28.112506       1 pv_controller.go:823] claim \"persistent-local-volumes-test-8575/pvc-p4ps5\" entered phase \"Bound\"\nI0718 20:02:28.368234       1 pvc_protection_controller.go:291] \"PVC is unused\" PVC=\"provisioning-7043/csi-hostpath7gbch\"\nI0718 20:02:28.376869       1 pv_controller.go:640] volume \"pvc-21f66de1-37b8-4329-a5b5-bd0271c6e945\" is released and reclaim policy \"Delete\" will be executed\nI0718 20:02:28.381397       1 pv_controller.go:879] volume \"pvc-21f66de1-37b8-4329-a5b5-bd0271c6e945\" entered phase \"Released\"\nI0718 20:02:28.393869       1 pv_controller.go:1341] isVolumeReleased[pvc-21f66de1-37b8-4329-a5b5-bd0271c6e945]: volume is released\nI0718 20:02:28.414613       1 pv_controller_base.go:505] deletion of claim \"provisioning-7043/csi-hostpath7gbch\" was already processed\nI0718 20:02:28.442325       1 namespace_controller.go:185] Namespace has been deleted provisioning-1806-9608\nI0718 20:02:29.133392       1 pvc_protection_controller.go:303] \"Pod uses PVC\" pod=\"ephemeral-599/inline-volume-tester-fg8zx\" PVC=\"ephemeral-599/inline-volume-tester-fg8zx-my-volume-0\"\nI0718 20:02:29.133419       1 pvc_protection_controller.go:181] \"Keeping PVC because it is being used\" PVC=\"ephemeral-599/inline-volume-tester-fg8zx-my-volume-0\"\nI0718 20:02:29.535737       1 pvc_protection_controller.go:291] \"PVC is unused\" PVC=\"ephemeral-599/inline-volume-tester-fg8zx-my-volume-0\"\nI0718 20:02:29.545768       1 garbagecollector.go:471] \"Processing object\" object=\"ephemeral-599/inline-volume-tester-fg8zx\" objectUID=121b3ea5-1324-4b58-87cc-328311442635 kind=\"Pod\" virtual=false\nI0718 20:02:29.549504       1 pv_controller.go:640] volume \"pvc-5ac74b95-2112-475c-9568-a2c5e6e5ef2b\" is released and reclaim policy \"Delete\" will be executed\nI0718 20:02:29.563741       1 garbagecollector.go:590] remove DeleteDependents finalizer for item [v1/Pod, namespace: ephemeral-599, name: inline-volume-tester-fg8zx, uid: 121b3ea5-1324-4b58-87cc-328311442635]\nI0718 20:02:29.563793       1 pv_controller.go:879] volume \"pvc-5ac74b95-2112-475c-9568-a2c5e6e5ef2b\" entered phase \"Released\"\nI0718 20:02:29.566900       1 pv_controller.go:1341] isVolumeReleased[pvc-5ac74b95-2112-475c-9568-a2c5e6e5ef2b]: volume is released\nI0718 20:02:29.588293       1 pv_controller_base.go:505] deletion of claim \"ephemeral-599/inline-volume-tester-fg8zx-my-volume-0\" was already processed\nI0718 20:02:29.984138       1 pvc_protection_controller.go:291] \"PVC is unused\" PVC=\"volume-6255/pvc-4xrbd\"\nI0718 20:02:29.997187       1 pv_controller.go:640] volume \"local-9cz82\" is released and reclaim policy \"Retain\" will be executed\nI0718 20:02:30.005171       1 pv_controller.go:879] volume \"local-9cz82\" entered phase \"Released\"\nI0718 20:02:30.026630       1 pv_controller_base.go:505] deletion of claim \"volume-6255/pvc-4xrbd\" was already processed\nI0718 20:02:30.739238       1 namespace_controller.go:185] Namespace has been deleted volume-7899\nE0718 20:02:30.890815       1 reflector.go:138] k8s.io/client-go/metadata/metadatainformer/informer.go:90: Failed to watch *v1.PartialObjectMetadata: failed to list *v1.PartialObjectMetadata: the server could not find the requested resource\nI0718 20:02:31.146136       1 namespace_controller.go:185] Namespace has been deleted webhook-3379-markers\nI0718 20:02:31.170265       1 namespace_controller.go:185] Namespace has been deleted webhook-3379\nI0718 20:02:31.314150       1 namespace_controller.go:185] Namespace has been deleted emptydir-7240\nI0718 20:02:31.539132       1 namespace_controller.go:185] Namespace has been deleted provisioning-3592\nE0718 20:02:31.560358       1 tokens_controller.go:262] error synchronizing serviceaccount kubectl-662/default: secrets \"default-token-m7k7w\" is forbidden: unable to create new content in namespace kubectl-662 because it is being terminated\nI0718 20:02:31.769066       1 namespace_controller.go:185] Namespace has been deleted dns-autoscaling-2257\nI0718 20:02:32.174308       1 namespace_controller.go:185] Namespace has been deleted persistent-local-volumes-test-448\nE0718 20:02:32.291722       1 tokens_controller.go:262] error synchronizing serviceaccount gc-7976/default: secrets \"default-token-f6kdz\" is forbidden: unable to create new content in namespace gc-7976 because it is being terminated\nI0718 20:02:32.347841       1 pv_controller.go:879] volume \"local-pvnlx44\" entered phase \"Available\"\nI0718 20:02:32.377154       1 pv_controller.go:930] claim \"persistent-local-volumes-test-8656/pvc-q52dg\" bound to volume \"local-pvnlx44\"\nI0718 20:02:32.390760       1 pv_controller.go:879] volume \"local-pvnlx44\" entered phase \"Bound\"\nI0718 20:02:32.390782       1 pv_controller.go:982] volume \"local-pvnlx44\" bound to claim \"persistent-local-volumes-test-8656/pvc-q52dg\"\nI0718 20:02:32.404356       1 pv_controller.go:823] claim \"persistent-local-volumes-test-8656/pvc-q52dg\" entered phase \"Bound\"\nE0718 20:02:32.425478       1 reflector.go:138] k8s.io/client-go/metadata/metadatainformer/informer.go:90: Failed to watch *v1.PartialObjectMetadata: failed to list *v1.PartialObjectMetadata: the server could not find the requested resource\nI0718 20:02:32.455340       1 reconciler.go:219] attacherDetacher.DetachVolume started for volume \"pvc-21f66de1-37b8-4329-a5b5-bd0271c6e945\" (UniqueName: \"kubernetes.io/csi/csi-hostpath-provisioning-7043^066c3db9-e803-11eb-8130-6692668a95ce\") on node \"10.116.0.7\" \nI0718 20:02:32.458622       1 operation_generator.go:1483] Verified volume is safe to detach for volume \"pvc-21f66de1-37b8-4329-a5b5-bd0271c6e945\" (UniqueName: \"kubernetes.io/csi/csi-hostpath-provisioning-7043^066c3db9-e803-11eb-8130-6692668a95ce\") on node \"10.116.0.7\" \nI0718 20:02:32.465301       1 reconciler.go:219] attacherDetacher.DetachVolume started for volume \"pvc-5ac74b95-2112-475c-9568-a2c5e6e5ef2b\" (UniqueName: \"kubernetes.io/csi/csi-hostpath-ephemeral-599^f1d72751-e802-11eb-8a90-6ecc6361dcfa\") on node \"10.116.0.7\" \nI0718 20:02:32.468760       1 operation_generator.go:1483] Verified volume is safe to detach for volume \"pvc-5ac74b95-2112-475c-9568-a2c5e6e5ef2b\" (UniqueName: \"kubernetes.io/csi/csi-hostpath-ephemeral-599^f1d72751-e802-11eb-8a90-6ecc6361dcfa\") on node \"10.116.0.7\" \nE0718 20:02:32.735694       1 tokens_controller.go:262] error synchronizing serviceaccount secrets-4740/default: secrets \"default-token-5brf4\" is forbidden: unable to create new content in namespace secrets-4740 because it is being terminated\nI0718 20:02:33.040749       1 operation_generator.go:483] DetachVolume.Detach succeeded for volume \"pvc-5ac74b95-2112-475c-9568-a2c5e6e5ef2b\" (UniqueName: \"kubernetes.io/csi/csi-hostpath-ephemeral-599^f1d72751-e802-11eb-8a90-6ecc6361dcfa\") on node \"10.116.0.7\" \nI0718 20:02:33.043893       1 operation_generator.go:483] DetachVolume.Detach succeeded for volume \"pvc-21f66de1-37b8-4329-a5b5-bd0271c6e945\" (UniqueName: \"kubernetes.io/csi/csi-hostpath-provisioning-7043^066c3db9-e803-11eb-8130-6692668a95ce\") on node \"10.116.0.7\" \nE0718 20:02:34.515128       1 pv_controller.go:1452] error finding provisioning plugin for claim provisioning-3398/pvc-r869k: storageclass.storage.k8s.io \"provisioning-3398\" not found\nI0718 20:02:34.515162       1 event.go:291] \"Event occurred\" object=\"provisioning-3398/pvc-r869k\" kind=\"PersistentVolumeClaim\" apiVersion=\"v1\" type=\"Warning\" reason=\"ProvisioningFailed\" message=\"storageclass.storage.k8s.io \\\"provisioning-3398\\\" not found\"\nI0718 20:02:34.556979       1 pv_controller.go:879] volume \"local-jmrl9\" entered phase \"Available\"\nE0718 20:02:35.260389       1 reflector.go:138] k8s.io/client-go/metadata/metadatainformer/informer.go:90: Failed to watch *v1.PartialObjectMetadata: failed to list *v1.PartialObjectMetadata: the server could not find the requested resource\nI0718 20:02:35.518819       1 garbagecollector.go:471] \"Processing object\" object=\"services-4562/externalsvc-vvx8f\" objectUID=69acf845-af20-48cb-b841-45f2df61fc9b kind=\"Pod\" virtual=false\nI0718 20:02:35.518904       1 garbagecollector.go:471] \"Processing object\" object=\"services-4562/externalsvc-ct777\" objectUID=0dadcc1b-3395-4c35-9fc4-d085a8ae14ad kind=\"Pod\" virtual=false\nI0718 20:02:35.525129       1 garbagecollector.go:580] \"Deleting object\" object=\"services-4562/externalsvc-vvx8f\" objectUID=69acf845-af20-48cb-b841-45f2df61fc9b kind=\"Pod\" propagationPolicy=Background\nI0718 20:02:35.525195       1 garbagecollector.go:580] \"Deleting object\" object=\"services-4562/externalsvc-ct777\" objectUID=0dadcc1b-3395-4c35-9fc4-d085a8ae14ad kind=\"Pod\" propagationPolicy=Background\nE0718 20:02:35.528584       1 tokens_controller.go:262] error synchronizing serviceaccount ephemeral-599/default: secrets \"default-token-qs8h2\" is forbidden: unable to create new content in namespace ephemeral-599 because it is being terminated\nI0718 20:02:35.682519       1 event.go:291] \"Event occurred\" object=\"deployment-4244/webserver\" kind=\"Deployment\" apiVersion=\"apps/v1\" type=\"Normal\" reason=\"ScalingReplicaSet\" message=\"Scaled up replica set webserver-847dcfb7fb to 6\"\nI0718 20:02:35.685816       1 replica_set.go:559] \"Too few replicas\" replicaSet=\"deployment-4244/webserver-847dcfb7fb\" need=6 creating=6\nI0718 20:02:35.693635       1 event.go:291] \"Event occurred\" object=\"deployment-4244/webserver-847dcfb7fb\" kind=\"ReplicaSet\" apiVersion=\"apps/v1\" type=\"Normal\" reason=\"SuccessfulCreate\" message=\"Created pod: webserver-847dcfb7fb-2wqw5\"\nI0718 20:02:35.695368       1 deployment_controller.go:490] \"Error syncing deployment\" deployment=\"deployment-4244/webserver\" err=\"Operation cannot be fulfilled on deployments.apps \\\"webserver\\\": the object has been modified; please apply your changes to the latest version and try again\"\nI0718 20:02:35.702725       1 event.go:291] \"Event occurred\" object=\"deployment-4244/webserver-847dcfb7fb\" kind=\"ReplicaSet\" apiVersion=\"apps/v1\" type=\"Normal\" reason=\"SuccessfulCreate\" message=\"Created pod: webserver-847dcfb7fb-j676f\"\nI0718 20:02:35.702853       1 event.go:291] \"Event occurred\" object=\"deployment-4244/webserver-847dcfb7fb\" kind=\"ReplicaSet\" apiVersion=\"apps/v1\" type=\"Normal\" reason=\"SuccessfulCreate\" message=\"Created pod: webserver-847dcfb7fb-fztl5\"\nI0718 20:02:35.710674       1 event.go:291] \"Event occurred\" object=\"deployment-4244/webserver-847dcfb7fb\" kind=\"ReplicaSet\" apiVersion=\"apps/v1\" type=\"Normal\" reason=\"SuccessfulCreate\" message=\"Created pod: webserver-847dcfb7fb-8fglq\"\nI0718 20:02:35.711123       1 event.go:291] \"Event occurred\" object=\"deployment-4244/webserver-847dcfb7fb\" kind=\"ReplicaSet\" apiVersion=\"apps/v1\" type=\"Normal\" reason=\"SuccessfulCreate\" message=\"Created pod: webserver-847dcfb7fb-ksg4n\"\nI0718 20:02:35.712677       1 event.go:291] \"Event occurred\" object=\"deployment-4244/webserver-847dcfb7fb\" kind=\"ReplicaSet\" apiVersion=\"apps/v1\" type=\"Normal\" reason=\"SuccessfulCreate\" message=\"Created pod: webserver-847dcfb7fb-zjtq6\"\nE0718 20:02:35.774916       1 namespace_controller.go:162] deletion of namespace volume-6255 failed: unable to retrieve the complete list of server APIs: crd-publish-openapi-test-multi-ver.example.com/v2: the server could not find the requested resource, crd-publish-openapi-test-multi-ver.example.com/v3: the server could not find the requested resource\nE0718 20:02:35.811401       1 namespace_controller.go:162] deletion of namespace ephemeral-599 failed: unable to retrieve the complete list of server APIs: crd-publish-openapi-test-multi-ver.example.com/v2: the server could not find the requested resource, crd-publish-openapi-test-multi-ver.example.com/v3: the server could not find the requested resource\nI0718 20:02:35.909178       1 namespace_controller.go:185] Namespace has been deleted provisioning-4462-6849\nE0718 20:02:35.911567       1 reflector.go:138] k8s.io/client-go/metadata/metadatainformer/informer.go:90: Failed to watch *v1.PartialObjectMetadata: failed to list *v1.PartialObjectMetadata: the server could not find the requested resource\nE0718 20:02:35.938564       1 namespace_controller.go:162] deletion of namespace volume-6255 failed: unable to retrieve the complete list of server APIs: crd-publish-openapi-test-multi-ver.example.com/v2: the server could not find the requested resource, crd-publish-openapi-test-multi-ver.example.com/v3: the server could not find the requested resource\nE0718 20:02:35.984017       1 namespace_controller.go:162] deletion of namespace ephemeral-599 failed: unable to retrieve the complete list of server APIs: crd-publish-openapi-test-multi-ver.example.com/v2: the server could not find the requested resource, crd-publish-openapi-test-multi-ver.example.com/v3: the server could not find the requested resource\nE0718 20:02:36.104349       1 namespace_controller.go:162] deletion of namespace volume-6255 failed: unable to retrieve the complete list of server APIs: crd-publish-openapi-test-multi-ver.example.com/v2: the server could not find the requested resource, crd-publish-openapi-test-multi-ver.example.com/v3: the server could not find the requested resource\nE0718 20:02:36.148274       1 namespace_controller.go:162] deletion of namespace ephemeral-599 failed: unable to retrieve the complete list of server APIs: crd-publish-openapi-test-multi-ver.example.com/v2: the server could not find the requested resource, crd-publish-openapi-test-multi-ver.example.com/v3: the server could not find the requested resource\nE0718 20:02:36.284350       1 namespace_controller.go:162] deletion of namespace volume-6255 failed: unable to retrieve the complete list of server APIs: crd-publish-openapi-test-multi-ver.example.com/v2: the server could not find the requested resource, crd-publish-openapi-test-multi-ver.example.com/v3: the server could not find the requested resource\nE0718 20:02:36.332964       1 namespace_controller.go:162] deletion of namespace ephemeral-599 failed: unable to retrieve the complete list of server APIs: crd-publish-openapi-test-multi-ver.example.com/v2: the server could not find the requested resource, crd-publish-openapi-test-multi-ver.example.com/v3: the server could not find the requested resource\nE0718 20:02:36.482151       1 namespace_controller.go:162] deletion of namespace volume-6255 failed: unable to retrieve the complete list of server APIs: crd-publish-openapi-test-multi-ver.example.com/v2: the server could not find the requested resource, crd-publish-openapi-test-multi-ver.example.com/v3: the server could not find the requested resource\nE0718 20:02:36.524885       1 namespace_controller.go:162] deletion of namespace ephemeral-599 failed: unable to retrieve the complete list of server APIs: crd-publish-openapi-test-multi-ver.example.com/v2: the server could not find the requested resource, crd-publish-openapi-test-multi-ver.example.com/v3: the server could not find the requested resource\nI0718 20:02:36.569433       1 pv_controller.go:930] claim \"provisioning-3398/pvc-r869k\" bound to volume \"local-jmrl9\"\nI0718 20:02:36.584149       1 pv_controller.go:879] volume \"local-jmrl9\" entered phase \"Bound\"\nI0718 20:02:36.584175       1 pv_controller.go:982] volume \"local-jmrl9\" bound to claim \"provisioning-3398/pvc-r869k\"\nI0718 20:02:36.595222       1 pv_controller.go:823] claim \"provisioning-3398/pvc-r869k\" entered phase \"Bound\"\nI0718 20:02:36.634360       1 namespace_controller.go:185] Namespace has been deleted kubectl-662\nE0718 20:02:36.723254       1 namespace_controller.go:162] deletion of namespace volume-6255 failed: unable to retrieve the complete list of server APIs: crd-publish-openapi-test-multi-ver.example.com/v2: the server could not find the requested resource, crd-publish-openapi-test-multi-ver.example.com/v3: the server could not find the requested resource\nE0718 20:02:36.971194       1 namespace_controller.go:162] deletion of namespace ephemeral-599 failed: unable to retrieve the complete list of server APIs: crd-publish-openapi-test-multi-ver.example.com/v2: the server could not find the requested resource, crd-publish-openapi-test-multi-ver.example.com/v3: the server could not find the requested resource\nE0718 20:02:37.086229       1 namespace_controller.go:162] deletion of namespace volume-6255 failed: unable to retrieve the complete list of server APIs: crd-publish-openapi-test-multi-ver.example.com/v2: the server could not find the requested resource, crd-publish-openapi-test-multi-ver.example.com/v3: the server could not find the requested resource\nI0718 20:02:37.217430       1 namespace_controller.go:185] Namespace has been deleted containers-8694\nE0718 20:02:37.302339       1 namespace_controller.go:162] deletion of namespace ephemeral-599 failed: unable to retrieve the complete list of server APIs: crd-publish-openapi-test-multi-ver.example.com/v2: the server could not find the requested resource, crd-publish-openapi-test-multi-ver.example.com/v3: the server could not find the requested resource\nE0718 20:02:37.304500       1 pv_controller.go:1452] error finding provisioning plugin for claim provisioning-9406/pvc-c4v95: storageclass.storage.k8s.io \"provisioning-9406\" not found\nI0718 20:02:37.304594       1 event.go:291] \"Event occurred\" object=\"provisioning-9406/pvc-c4v95\" kind=\"PersistentVolumeClaim\" apiVersion=\"v1\" type=\"Warning\" reason=\"ProvisioningFailed\" message=\"storageclass.storage.k8s.io \\\"provisioning-9406\\\" not found\"\nI0718 20:02:37.341543       1 pv_controller.go:879] volume \"local-5jzbn\" entered phase \"Available\"\nI0718 20:02:37.384805       1 namespace_controller.go:185] Namespace has been deleted gc-7976\nE0718 20:02:37.616400       1 namespace_controller.go:162] deletion of namespace volume-6255 failed: unable to retrieve the complete list of server APIs: crd-publish-openapi-test-multi-ver.example.com/v2: the server could not find the requested resource, crd-publish-openapi-test-multi-ver.example.com/v3: the server could not find the requested resource\nI0718 20:02:37.811907       1 replica_set.go:559] \"Too few replicas\" replicaSet=\"deployment-4244/webserver-847dcfb7fb\" need=7 creating=1\nI0718 20:02:37.812032       1 event.go:291] \"Event occurred\" object=\"deployment-4244/webserver\" kind=\"Deployment\" apiVersion=\"apps/v1\" type=\"Normal\" reason=\"ScalingReplicaSet\" message=\"Scaled up replica set webserver-847dcfb7fb to 7\"\nI0718 20:02:37.830158       1 event.go:291] \"Event occurred\" object=\"deployment-4244/webserver-847dcfb7fb\" kind=\"ReplicaSet\" apiVersion=\"apps/v1\" type=\"Normal\" reason=\"SuccessfulCreate\" message=\"Created pod: webserver-847dcfb7fb-cljkg\"\nE0718 20:02:37.850859       1 namespace_controller.go:162] deletion of namespace ephemeral-599 failed: unable to retrieve the complete list of server APIs: crd-publish-openapi-test-multi-ver.example.com/v2: the server could not find the requested resource, crd-publish-openapi-test-multi-ver.example.com/v3: the server could not find the requested resource\nI0718 20:02:37.898067       1 namespace_controller.go:185] Namespace has been deleted secrets-4740\nE0718 20:02:37.953642       1 namespace_controller.go:162] deletion of namespace secrets-7800 failed: unable to retrieve the complete list of server APIs: crd-publish-openapi-test-multi-ver.example.com/v2: the server could not find the requested resource, crd-publish-openapi-test-multi-ver.example.com/v3: the server could not find the requested resource\nE0718 20:02:38.124069       1 namespace_controller.go:162] deletion of namespace secrets-7800 failed: unable to retrieve the complete list of server APIs: crd-publish-openapi-test-multi-ver.example.com/v2: the server could not find the requested resource, crd-publish-openapi-test-multi-ver.example.com/v3: the server could not find the requested resource\nI0718 20:02:38.287744       1 replica_set.go:559] \"Too few replicas\" replicaSet=\"deployment-4244/webserver-6584b976d5\" need=2 creating=2\nI0718 20:02:38.287790       1 event.go:291] \"Event occurred\" object=\"deployment-4244/webserver\" kind=\"Deployment\" apiVersion=\"apps/v1\" type=\"Normal\" reason=\"ScalingReplicaSet\" message=\"Scaled up replica set webserver-6584b976d5 to 2\"\nI0718 20:02:38.295443       1 event.go:291] \"Event occurred\" object=\"deployment-4244/webserver-6584b976d5\" kind=\"ReplicaSet\" apiVersion=\"apps/v1\" type=\"Normal\" reason=\"SuccessfulCreate\" message=\"Created pod: webserver-6584b976d5-lf6r4\"\nI0718 20:02:38.304494       1 replica_set.go:595] \"Too many replicas\" replicaSet=\"deployment-4244/webserver-847dcfb7fb\" need=6 deleting=1\nI0718 20:02:38.304525       1 replica_set.go:223] \"Found related ReplicaSets\" replicaSet=\"deployment-4244/webserver-847dcfb7fb\" relatedReplicaSets=[webserver-6584b976d5 webserver-847dcfb7fb]\nI0718 20:02:38.304579       1 controller_utils.go:602] \"Deleting pod\" controller=\"webserver-847dcfb7fb\" pod=\"deployment-4244/webserver-847dcfb7fb-j676f\"\nI0718 20:02:38.304738       1 event.go:291] \"Event occurred\" object=\"deployment-4244/webserver-6584b976d5\" kind=\"ReplicaSet\" apiVersion=\"apps/v1\" type=\"Normal\" reason=\"SuccessfulCreate\" message=\"Created pod: webserver-6584b976d5-j4tn2\"\nI0718 20:02:38.304767       1 event.go:291] \"Event occurred\" object=\"deployment-4244/webserver\" kind=\"Deployment\" apiVersion=\"apps/v1\" type=\"Normal\" reason=\"ScalingReplicaSet\" message=\"Scaled down replica set webserver-847dcfb7fb to 6\"\nI0718 20:02:38.325468       1 event.go:291] \"Event occurred\" object=\"deployment-4244/webserver-847dcfb7fb\" kind=\"ReplicaSet\" apiVersion=\"apps/v1\" type=\"Normal\" reason=\"SuccessfulDelete\" message=\"Deleted pod: webserver-847dcfb7fb-j676f\"\nI0718 20:02:38.326696       1 deployment_controller.go:490] \"Error syncing deployment\" deployment=\"deployment-4244/webserver\" err=\"Operation cannot be fulfilled on deployments.apps \\\"webserver\\\": the object has been modified; please apply your changes to the latest version and try again\"\nI0718 20:02:38.334443       1 replica_set.go:559] \"Too few replicas\" replicaSet=\"deployment-4244/webserver-6584b976d5\" need=3 creating=1\nI0718 20:02:38.334621       1 event.go:291] \"Event occurred\" object=\"deployment-4244/webserver\" kind=\"Deployment\" apiVersion=\"apps/v1\" type=\"Normal\" reason=\"ScalingReplicaSet\" message=\"Scaled up replica set webserver-6584b976d5 to 3\"\nI0718 20:02:38.341403       1 event.go:291] \"Event occurred\" object=\"deployment-4244/webserver-6584b976d5\" kind=\"ReplicaSet\" apiVersion=\"apps/v1\" type=\"Normal\" reason=\"SuccessfulCreate\" message=\"Created pod: webserver-6584b976d5-xk67h\"\nE0718 20:02:38.361410       1 namespace_controller.go:162] deletion of namespace secrets-7800 failed: unable to retrieve the complete list of server APIs: crd-publish-openapi-test-multi-ver.example.com/v2: the server could not find the requested resource, crd-publish-openapi-test-multi-ver.example.com/v3: the server could not find the requested resource\nE0718 20:02:38.465736       1 namespace_controller.go:162] deletion of namespace volume-6255 failed: unable to retrieve the complete list of server APIs: crd-publish-openapi-test-multi-ver.example.com/v2: the server could not find the requested resource, crd-publish-openapi-test-multi-ver.example.com/v3: the server could not find the requested resource\nE0718 20:02:38.555057       1 namespace_controller.go:162] deletion of namespace secrets-7800 failed: unable to retrieve the complete list of server APIs: crd-publish-openapi-test-multi-ver.example.com/v2: the server could not find the requested resource, crd-publish-openapi-test-multi-ver.example.com/v3: the server could not find the requested resource\nE0718 20:02:38.586961       1 tokens_controller.go:262] error synchronizing serviceaccount provisioning-7043/default: secrets \"default-token-j4hjb\" is forbidden: unable to create new content in namespace provisioning-7043 because it is being terminated\nE0718 20:02:38.683660       1 namespace_controller.go:162] deletion of namespace ephemeral-599 failed: unable to retrieve the complete list of server APIs: crd-publish-openapi-test-multi-ver.example.com/v2: the server could not find the requested resource, crd-publish-openapi-test-multi-ver.example.com/v3: the server could not find the requested resource\nE0718 20:02:38.759097       1 namespace_controller.go:162] deletion of namespace secrets-7800 failed: unable to retrieve the complete list of server APIs: crd-publish-openapi-test-multi-ver.example.com/v2: the server could not find the requested resource, crd-publish-openapi-test-multi-ver.example.com/v3: the server could not find the requested resource\nE0718 20:02:38.801137       1 namespace_controller.go:162] deletion of namespace provisioning-7043 failed: unable to retrieve the complete list of server APIs: crd-publish-openapi-test-multi-ver.example.com/v2: the server could not find the requested resource, crd-publish-openapi-test-multi-ver.example.com/v3: the server could not find the requested resource\nE0718 20:02:38.960576       1 namespace_controller.go:162] deletion of namespace provisioning-7043 failed: unable to retrieve the complete list of server APIs: crd-publish-openapi-test-multi-ver.example.com/v2: the server could not find the requested resource, crd-publish-openapi-test-multi-ver.example.com/v3: the server could not find the requested resource\nE0718 20:02:39.009032       1 namespace_controller.go:162] deletion of namespace secrets-7800 failed: unable to retrieve the complete list of server APIs: crd-publish-openapi-test-multi-ver.example.com/v2: the server could not find the requested resource, crd-publish-openapi-test-multi-ver.example.com/v3: the server could not find the requested resource\nE0718 20:02:39.145825       1 namespace_controller.go:162] deletion of namespace provisioning-7043 failed: unable to retrieve the complete list of server APIs: crd-publish-openapi-test-multi-ver.example.com/v2: the server could not find the requested resource, crd-publish-openapi-test-multi-ver.example.com/v3: the server could not find the requested resource\nE0718 20:02:39.405443       1 namespace_controller.go:162] deletion of namespace provisioning-7043 failed: unable to retrieve the complete list of server APIs: crd-publish-openapi-test-multi-ver.example.com/v2: the server could not find the requested resource, crd-publish-openapi-test-multi-ver.example.com/v3: the server could not find the requested resource\nE0718 20:02:39.411391       1 namespace_controller.go:162] deletion of namespace secrets-7800 failed: unable to retrieve the complete list of server APIs: crd-publish-openapi-test-multi-ver.example.com/v2: the server could not find the requested resource, crd-publish-openapi-test-multi-ver.example.com/v3: the server could not find the requested resource\nE0718 20:02:39.645148       1 namespace_controller.go:162] deletion of namespace provisioning-7043 failed: unable to retrieve the complete list of server APIs: crd-publish-openapi-test-multi-ver.example.com/v2: the server could not find the requested resource, crd-publish-openapi-test-multi-ver.example.com/v3: the server could not find the requested resource\nE0718 20:02:39.947362       1 namespace_controller.go:162] deletion of namespace secrets-7800 failed: unable to retrieve the complete list of server APIs: crd-publish-openapi-test-multi-ver.example.com/v2: the server could not find the requested resource, crd-publish-openapi-test-multi-ver.example.com/v3: the server could not find the requested resource\nE0718 20:02:40.099821       1 namespace_controller.go:162] deletion of namespace provisioning-7043 failed: unable to retrieve the complete list of server APIs: crd-publish-openapi-test-multi-ver.example.com/v2: the server could not find the requested resource, crd-publish-openapi-test-multi-ver.example.com/v3: the server could not find the requested resource\nE0718 20:02:40.153487       1 namespace_controller.go:162] deletion of namespace kubelet-test-4596 failed: unable to retrieve the complete list of server APIs: crd-publish-openapi-test-multi-ver.example.com/v2: the server could not find the requested resource, crd-publish-openapi-test-multi-ver.example.com/v3: the server could not find the requested resource\nE0718 20:02:40.187988       1 namespace_controller.go:162] deletion of namespace volume-6255 failed: unable to retrieve the complete list of server APIs: crd-publish-openapi-test-multi-ver.example.com/v2: the server could not find the requested resource, crd-publish-openapi-test-multi-ver.example.com/v3: the server could not find the requested resource\nE0718 20:02:40.255721       1 namespace_controller.go:162] deletion of namespace ephemeral-599 failed: unable to retrieve the complete list of server APIs: crd-publish-openapi-test-multi-ver.example.com/v2: the server could not find the requested resource, crd-publish-openapi-test-multi-ver.example.com/v3: the server could not find the requested resource\nI0718 20:02:40.421963       1 replica_set.go:595] \"Too many replicas\" replicaSet=\"deployment-4244/webserver-847dcfb7fb\" need=5 deleting=1\nI0718 20:02:40.422001       1 replica_set.go:223] \"Found related ReplicaSets\" replicaSet=\"deployment-4244/webserver-847dcfb7fb\" relatedReplicaSets=[webserver-847dcfb7fb webserver-6584b976d5]\nI0718 20:02:40.422065       1 event.go:291] \"Event occurred\" object=\"deployment-4244/webserver\" kind=\"Deployment\" apiVersion=\"apps/v1\" type=\"Normal\" reason=\"ScalingReplicaSet\" message=\"Scaled down replica set webserver-847dcfb7fb to 5\"\nI0718 20:02:40.422095       1 controller_utils.go:602] \"Deleting pod\" controller=\"webserver-847dcfb7fb\" pod=\"deployment-4244/webserver-847dcfb7fb-8fglq\"\nI0718 20:02:40.456310       1 event.go:291] \"Event occurred\" object=\"deployment-4244/webserver-847dcfb7fb\" kind=\"ReplicaSet\" apiVersion=\"apps/v1\" type=\"Normal\" reason=\"SuccessfulDelete\" message=\"Deleted pod: webserver-847dcfb7fb-8fglq\"\nI0718 20:02:40.459333       1 deployment_controller.go:490] \"Error syncing deployment\" deployment=\"deployment-4244/webserver\" err=\"Operation cannot be fulfilled on deployments.apps \\\"webserver\\\": the object has been modified; please apply your changes to the latest version and try again\"\nI0718 20:02:40.478402       1 deployment_controller.go:490] \"Error syncing deployment\" deployment=\"deployment-4244/webserver\" err=\"Operation cannot be fulfilled on deployments.apps \\\"webserver\\\": the object has been modified; please apply your changes to the latest version and try again\"\nI0718 20:02:40.484289       1 replica_set.go:595] \"Too many replicas\" replicaSet=\"deployment-4244/webserver-847dcfb7fb\" need=2 deleting=3\nI0718 20:02:40.484323       1 replica_set.go:223] \"Found related ReplicaSets\" replicaSet=\"deployment-4244/webserver-847dcfb7fb\" relatedReplicaSets=[webserver-847dcfb7fb webserver-6584b976d5 webserver-789dfdff45]\nI0718 20:02:40.484386       1 event.go:291] \"Event occurred\" object=\"deployment-4244/webserver\" kind=\"Deployment\" apiVersion=\"apps/v1\" type=\"Normal\" reason=\"ScalingReplicaSet\" message=\"Scaled down replica set webserver-847dcfb7fb to 2\"\nI0718 20:02:40.484406       1 controller_utils.go:602] \"Deleting pod\" controller=\"webserver-847dcfb7fb\" pod=\"deployment-4244/webserver-847dcfb7fb-2wqw5\"\nI0718 20:02:40.484431       1 controller_utils.go:602] \"Deleting pod\" controller=\"webserver-847dcfb7fb\" pod=\"deployment-4244/webserver-847dcfb7fb-ksg4n\"\nI0718 20:02:40.484421       1 controller_utils.go:602] \"Deleting pod\" controller=\"webserver-847dcfb7fb\" pod=\"deployment-4244/webserver-847dcfb7fb-zjtq6\"\nI0718 20:02:40.498824       1 event.go:291] \"Event occurred\" object=\"deployment-4244/webserver-847dcfb7fb\" kind=\"ReplicaSet\" apiVersion=\"apps/v1\" type=\"Normal\" reason=\"SuccessfulDelete\" message=\"Deleted pod: webserver-847dcfb7fb-zjtq6\"\nI0718 20:02:40.506433       1 replica_set.go:559] \"Too few replicas\" replicaSet=\"deployment-4244/webserver-789dfdff45\" need=3 creating=3\nI0718 20:02:40.506673       1 event.go:291] \"Event occurred\" object=\"deployment-4244/webserver\" kind=\"Deployment\" apiVersion=\"apps/v1\" type=\"Normal\" reason=\"ScalingReplicaSet\" message=\"Scaled up replica set webserver-789dfdff45 to 3\"\nI0718 20:02:40.506941       1 event.go:291] \"Event occurred\" object=\"deployment-4244/webserver-847dcfb7fb\" kind=\"ReplicaSet\" apiVersion=\"apps/v1\" type=\"Normal\" reason=\"SuccessfulDelete\" message=\"Deleted pod: webserver-847dcfb7fb-2wqw5\"\nI0718 20:02:40.510234       1 event.go:291] \"Event occurred\" object=\"deployment-4244/webserver-847dcfb7fb\" kind=\"ReplicaSet\" apiVersion=\"apps/v1\" type=\"Normal\" reason=\"SuccessfulDelete\" message=\"Deleted pod: webserver-847dcfb7fb-ksg4n\"\nI0718 20:02:40.515789       1 event.go:291] \"Event occurred\" object=\"deployment-4244/webserver-789dfdff45\" kind=\"ReplicaSet\" apiVersion=\"apps/v1\" type=\"Normal\" reason=\"SuccessfulCreate\" message=\"Created pod: webserver-789dfdff45-ntnbk\"\nI0718 20:02:40.524903       1 event.go:291] \"Event occurred\" object=\"deployment-4244/webserver-789dfdff45\" kind=\"ReplicaSet\" apiVersion=\"apps/v1\" type=\"Normal\" reason=\"SuccessfulCreate\" message=\"Created pod: webserver-789dfdff45-mqvsl\"\nI0718 20:02:40.526359       1 event.go:291] \"Event occurred\" object=\"deployment-4244/webserver-789dfdff45\" kind=\"ReplicaSet\" apiVersion=\"apps/v1\" type=\"Normal\" reason=\"SuccessfulCreate\" message=\"Created pod: webserver-789dfdff45-pqfk7\"\nI0718 20:02:40.552479       1 deployment_controller.go:490] \"Error syncing deployment\" deployment=\"deployment-4244/webserver\" err=\"Operation cannot be fulfilled on deployments.apps \\\"webserver\\\": the object has been modified; please apply your changes to the latest version and try again\"\nI0718 20:02:40.678266       1 event.go:291] \"Event occurred\" object=\"csi-mock-volumes-9933-3400/csi-mockplugin\" kind=\"StatefulSet\" apiVersion=\"apps/v1\" type=\"Normal\" reason=\"SuccessfulCreate\" message=\"create Pod csi-mockplugin-0 in StatefulSet csi-mockplugin successful\"\nI0718 20:02:40.705702       1 event.go:291] \"Event occurred\" object=\"csi-mock-volumes-9933-3400/csi-mockplugin-attacher\" kind=\"StatefulSet\" apiVersion=\"apps/v1\" type=\"Normal\" reason=\"SuccessfulCreate\" message=\"create Pod csi-mockplugin-attacher-0 in StatefulSet csi-mockplugin-attacher successful\"\nI0718 20:02:40.740285       1 event.go:291] \"Event occurred\" object=\"csi-mock-volumes-9933-3400/csi-mockplugin-resizer\" kind=\"StatefulSet\" apiVersion=\"apps/v1\" type=\"Normal\" reason=\"SuccessfulCreate\" message=\"create Pod csi-mockplugin-resizer-0 in StatefulSet csi-mockplugin-resizer successful\"\nE0718 20:02:40.968535       1 reflector.go:138] k8s.io/client-go/metadata/metadatainformer/informer.go:90: Failed to watch *v1.PartialObjectMetadata: failed to list *v1.PartialObjectMetadata: the server could not find the requested resource\nE0718 20:02:41.013496       1 tokens_controller.go:262] error synchronizing serviceaccount projected-7562/default: secrets \"default-token-6qm89\" is forbidden: unable to create new content in namespace projected-7562 because it is being terminated\nI0718 20:02:41.127763       1 pv_controller.go:879] volume \"local-pvdvlxc\" entered phase \"Available\"\nI0718 20:02:41.157733       1 pv_controller.go:930] claim \"persistent-local-volumes-test-2875/pvc-n9rch\" bound to volume \"local-pvdvlxc\"\nI0718 20:02:41.168407       1 pv_controller.go:879] volume \"local-pvdvlxc\" entered phase \"Bound\"\nI0718 20:02:41.168437       1 pv_controller.go:982] volume \"local-pvdvlxc\" bound to claim \"persistent-local-volumes-test-2875/pvc-n9rch\"\nI0718 20:02:41.181235       1 pv_controller.go:823] claim \"persistent-local-volumes-test-2875/pvc-n9rch\" entered phase \"Bound\"\nI0718 20:02:41.696687       1 pv_controller.go:879] volume \"local-pvzszt6\" entered phase \"Available\"\nI0718 20:02:41.729773       1 pv_controller.go:930] claim \"persistent-local-volumes-test-7688/pvc-798xd\" bound to volume \"local-pvzszt6\"\nI0718 20:02:41.741348       1 pv_controller.go:879] volume \"local-pvzszt6\" entered phase \"Bound\"\nI0718 20:02:41.741383       1 pv_controller.go:982] volume \"local-pvzszt6\" bound to claim \"persistent-local-volumes-test-7688/pvc-798xd\"\nI0718 20:02:41.755332       1 pv_controller.go:823] claim \"persistent-local-volumes-test-7688/pvc-798xd\" entered phase \"Bound\"\nI0718 20:02:41.891582       1 pvc_protection_controller.go:291] \"PVC is unused\" PVC=\"persistent-local-volumes-test-7688/pvc-798xd\"\nI0718 20:02:41.901378       1 pv_controller.go:640] volume \"local-pvzszt6\" is released and reclaim policy \"Retain\" will be executed\nI0718 20:02:41.910080       1 pv_controller.go:879] volume \"local-pvzszt6\" entered phase \"Released\"\nI0718 20:02:41.947656       1 pv_controller_base.go:505] deletion of claim \"persistent-local-volumes-test-7688/pvc-798xd\" was already processed\nI0718 20:02:42.340925       1 garbagecollector.go:471] \"Processing object\" object=\"provisioning-7043-7873/csi-hostpath-attacher-msrtl\" objectUID=d83c60c8-44f4-4dc5-9a61-25b52231e9c7 kind=\"EndpointSlice\" virtual=false\nI0718 20:02:42.351908       1 garbagecollector.go:580] \"Deleting object\" object=\"provisioning-7043-7873/csi-hostpath-attacher-msrtl\" objectUID=d83c60c8-44f4-4dc5-9a61-25b52231e9c7 kind=\"EndpointSlice\" propagationPolicy=Background\nI0718 20:02:42.396337       1 garbagecollector.go:471] \"Processing object\" object=\"provisioning-7043-7873/csi-hostpath-attacher-77d7d6bd8b\" objectUID=f3c075c9-dad7-4241-8fe2-91c98758f486 kind=\"ControllerRevision\" virtual=false\nI0718 20:02:42.396404       1 stateful_set.go:419] StatefulSet has been deleted provisioning-7043-7873/csi-hostpath-attacher\nI0718 20:02:42.396457       1 garbagecollector.go:471] \"Processing object\" object=\"provisioning-7043-7873/csi-hostpath-attacher-0\" objectUID=39dc118f-16a1-4ade-b5cf-d423a2cd8a0d kind=\"Pod\" virtual=false\nI0718 20:02:42.398773       1 garbagecollector.go:580] \"Deleting object\" object=\"provisioning-7043-7873/csi-hostpath-attacher-0\" objectUID=39dc118f-16a1-4ade-b5cf-d423a2cd8a0d kind=\"Pod\" propagationPolicy=Background\nI0718 20:02:42.399648       1 garbagecollector.go:580] \"Deleting object\" object=\"provisioning-7043-7873/csi-hostpath-attacher-77d7d6bd8b\" objectUID=f3c075c9-dad7-4241-8fe2-91c98758f486 kind=\"ControllerRevision\" propagationPolicy=Background\nI0718 20:02:42.467498       1 garbagecollector.go:471] \"Processing object\" object=\"provisioning-7043-7873/csi-hostpathplugin-dpr6b\" objectUID=72801da5-efb0-4916-bd4c-bdcaafa5795f kind=\"EndpointSlice\" virtual=false\nI0718 20:02:42.475282       1 garbagecollector.go:580] \"Deleting object\" object=\"provisioning-7043-7873/csi-hostpathplugin-dpr6b\" objectUID=72801da5-efb0-4916-bd4c-bdcaafa5795f kind=\"EndpointSlice\" propagationPolicy=Background\nI0718 20:02:42.518331       1 garbagecollector.go:471] \"Processing object\" object=\"provisioning-7043-7873/csi-hostpathplugin-647494c88b\" objectUID=4b2e19ae-47ff-4035-9f39-2f5e420e971a kind=\"ControllerRevision\" virtual=false\nI0718 20:02:42.518341       1 stateful_set.go:419] StatefulSet has been deleted provisioning-7043-7873/csi-hostpathplugin\nI0718 20:02:42.518476       1 garbagecollector.go:471] \"Processing object\" object=\"provisioning-7043-7873/csi-hostpathplugin-0\" objectUID=e1e102a0-0c81-43f3-9a01-7c5381a322ee kind=\"Pod\" virtual=false\nI0718 20:02:42.520606       1 garbagecollector.go:580] \"Deleting object\" object=\"provisioning-7043-7873/csi-hostpathplugin-647494c88b\" objectUID=4b2e19ae-47ff-4035-9f39-2f5e420e971a kind=\"ControllerRevision\" propagationPolicy=Background\nI0718 20:02:42.521583       1 garbagecollector.go:580] \"Deleting object\" object=\"provisioning-7043-7873/csi-hostpathplugin-0\" objectUID=e1e102a0-0c81-43f3-9a01-7c5381a322ee kind=\"Pod\" propagationPolicy=Background\nI0718 20:02:42.557742       1 replica_set.go:595] \"Too many replicas\" replicaSet=\"deployment-4244/webserver-6584b976d5\" need=2 deleting=1\nI0718 20:02:42.557782       1 replica_set.go:223] \"Found related ReplicaSets\" replicaSet=\"deployment-4244/webserver-6584b976d5\" relatedReplicaSets=[webserver-847dcfb7fb webserver-6584b976d5 webserver-789dfdff45]\nI0718 20:02:42.557791       1 garbagecollector.go:471] \"Processing object\" object=\"provisioning-7043-7873/csi-hostpath-provisioner-d6vpg\" objectUID=b79e59fc-a46c-42dc-bda5-8c3b5b319fe9 kind=\"EndpointSlice\" virtual=false\nI0718 20:02:42.557752       1 event.go:291] \"Event occurred\" object=\"deployment-4244/webserver\" kind=\"Deployment\" apiVersion=\"apps/v1\" type=\"Normal\" reason=\"ScalingReplicaSet\" message=\"Scaled down replica set webserver-6584b976d5 to 2\"\nI0718 20:02:42.557874       1 controller_utils.go:602] \"Deleting pod\" controller=\"webserver-6584b976d5\" pod=\"deployment-4244/webserver-6584b976d5-xk67h\"\nI0718 20:02:42.563551       1 garbagecollector.go:580] \"Deleting object\" object=\"provisioning-7043-7873/csi-hostpath-provisioner-d6vpg\" objectUID=b79e59fc-a46c-42dc-bda5-8c3b5b319fe9 kind=\"EndpointSlice\" propagationPolicy=Background\nI0718 20:02:42.586839       1 event.go:291] \"Event occurred\" object=\"deployment-4244/webserver-6584b976d5\" kind=\"ReplicaSet\" apiVersion=\"apps/v1\" type=\"Normal\" reason=\"SuccessfulDelete\" message=\"Deleted pod: webserver-6584b976d5-xk67h\"\nI0718 20:02:42.618109       1 garbagecollector.go:471] \"Processing object\" object=\"provisioning-7043-7873/csi-hostpath-provisioner-7bb56f8868\" objectUID=6327e986-2b3f-4f1d-bd07-a2e5f4beec35 kind=\"ControllerRevision\" virtual=false\nI0718 20:02:42.618251       1 stateful_set.go:419] StatefulSet has been deleted provisioning-7043-7873/csi-hostpath-provisioner\nI0718 20:02:42.618295       1 garbagecollector.go:471] \"Processing object\" object=\"provisioning-7043-7873/csi-hostpath-provisioner-0\" objectUID=ad41cea3-b472-4cc2-8f1a-d2a0d25909af kind=\"Pod\" virtual=false\nI0718 20:02:42.620337       1 garbagecollector.go:580] \"Deleting object\" object=\"provisioning-7043-7873/csi-hostpath-provisioner-7bb56f8868\" objectUID=6327e986-2b3f-4f1d-bd07-a2e5f4beec35 kind=\"ControllerRevision\" propagationPolicy=Background\nI0718 20:02:42.621140       1 garbagecollector.go:580] \"Deleting object\" object=\"provisioning-7043-7873/csi-hostpath-provisioner-0\" objectUID=ad41cea3-b472-4cc2-8f1a-d2a0d25909af kind=\"Pod\" propagationPolicy=Background\nI0718 20:02:42.653576       1 garbagecollector.go:471] \"Processing object\" object=\"provisioning-7043-7873/csi-hostpath-resizer-tb6hp\" objectUID=b671b761-94e3-4dee-a27f-be9c68cd2362 kind=\"EndpointSlice\" virtual=false\nI0718 20:02:42.660828       1 garbagecollector.go:580] \"Deleting object\" object=\"provisioning-7043-7873/csi-hostpath-resizer-tb6hp\" objectUID=b671b761-94e3-4dee-a27f-be9c68cd2362 kind=\"EndpointSlice\" propagationPolicy=Background\nI0718 20:02:42.705526       1 garbagecollector.go:471] \"Processing object\" object=\"provisioning-7043-7873/csi-hostpath-resizer-5d5957f9cf\" objectUID=d73692cc-0e8c-4af2-8f7d-87d52d86bf8a kind=\"ControllerRevision\" virtual=false\nI0718 20:02:42.705673       1 stateful_set.go:419] StatefulSet has been deleted provisioning-7043-7873/csi-hostpath-resizer\nI0718 20:02:42.705704       1 garbagecollector.go:471] \"Processing object\" object=\"provisioning-7043-7873/csi-hostpath-resizer-0\" objectUID=f6eee4b6-4fab-4932-8df1-f38caae8f370 kind=\"Pod\" virtual=false\nI0718 20:02:42.707634       1 garbagecollector.go:580] \"Deleting object\" object=\"provisioning-7043-7873/csi-hostpath-resizer-5d5957f9cf\" objectUID=d73692cc-0e8c-4af2-8f7d-87d52d86bf8a kind=\"ControllerRevision\" propagationPolicy=Background\nI0718 20:02:42.708441       1 garbagecollector.go:580] \"Deleting object\" object=\"provisioning-7043-7873/csi-hostpath-resizer-0\" objectUID=f6eee4b6-4fab-4932-8df1-f38caae8f370 kind=\"Pod\" propagationPolicy=Background\nI0718 20:02:42.740743       1 garbagecollector.go:471] \"Processing object\" object=\"provisioning-7043-7873/csi-hostpath-snapshotter-7b44x\" objectUID=fc3ca78d-14c3-4c79-b374-fca9cd2216c1 kind=\"EndpointSlice\" virtual=false\nI0718 20:02:42.743123       1 garbagecollector.go:580] \"Deleting object\" object=\"provisioning-7043-7873/csi-hostpath-snapshotter-7b44x\" objectUID=fc3ca78d-14c3-4c79-b374-fca9cd2216c1 kind=\"EndpointSlice\" propagationPolicy=Background\nI0718 20:02:42.795591       1 garbagecollector.go:471] \"Processing object\" object=\"provisioning-7043-7873/csi-hostpath-snapshotter-0\" objectUID=e2cfaf1e-7cfd-4412-a942-73a5d017da32 kind=\"Pod\" virtual=false\nI0718 20:02:42.795591       1 garbagecollector.go:471] \"Processing object\" object=\"provisioning-7043-7873/csi-hostpath-snapshotter-65dd87c6bc\" objectUID=e65729f8-2e7b-4653-ad17-0320489c50e1 kind=\"ControllerRevision\" virtual=false\nI0718 20:02:42.795592       1 stateful_set.go:419] StatefulSet has been deleted provisioning-7043-7873/csi-hostpath-snapshotter\nI0718 20:02:42.799901       1 garbagecollector.go:580] \"Deleting object\" object=\"provisioning-7043-7873/csi-hostpath-snapshotter-65dd87c6bc\" objectUID=e65729f8-2e7b-4653-ad17-0320489c50e1 kind=\"ControllerRevision\" propagationPolicy=Background\nI0718 20:02:42.799932       1 garbagecollector.go:580] \"Deleting object\" object=\"provisioning-7043-7873/csi-hostpath-snapshotter-0\" objectUID=e2cfaf1e-7cfd-4412-a942-73a5d017da32 kind=\"Pod\" propagationPolicy=Background\nE0718 20:02:43.043710       1 reflector.go:138] k8s.io/client-go/metadata/metadatainformer/informer.go:90: Failed to watch *v1.PartialObjectMetadata: failed to list *v1.PartialObjectMetadata: the server could not find the requested resource\nI0718 20:02:43.490385       1 pvc_protection_controller.go:303] \"Pod uses PVC\" pod=\"persistent-local-volumes-test-8656/pod-b15f46fd-7174-4b0a-b2ad-0fbc0f0b3578\" PVC=\"persistent-local-volumes-test-8656/pvc-q52dg\"\nI0718 20:02:43.490407       1 pvc_protection_controller.go:181] \"Keeping PVC because it is being used\" PVC=\"persistent-local-volumes-test-8656/pvc-q52dg\"\nI0718 20:02:43.516852       1 garbagecollector.go:471] \"Processing object\" object=\"services-8253/nodeport-reuse-tdm7w\" objectUID=eab58957-50a3-4c5b-94cf-960423ad8c21 kind=\"EndpointSlice\" virtual=false\nI0718 20:02:43.519289       1 garbagecollector.go:580] \"Deleting object\" object=\"services-8253/nodeport-reuse-tdm7w\" objectUID=eab58957-50a3-4c5b-94cf-960423ad8c21 kind=\"EndpointSlice\" propagationPolicy=Background\nI0718 20:02:43.905627       1 replica_set.go:559] \"Too few replicas\" replicaSet=\"deployment-4244/webserver-6584b976d5\" need=2 creating=1\nI0718 20:02:43.911587       1 event.go:291] \"Event occurred\" object=\"deployment-4244/webserver-6584b976d5\" kind=\"ReplicaSet\" apiVersion=\"apps/v1\" type=\"Normal\" reason=\"SuccessfulCreate\" message=\"Created pod: webserver-6584b976d5-52nlw\"\nI0718 20:02:43.941570       1 replica_set.go:559] \"Too few replicas\" replicaSet=\"deployment-4244/webserver-789dfdff45\" need=3 creating=1\nI0718 20:02:43.948088       1 event.go:291] \"Event occurred\" object=\"deployment-4244/webserver-789dfdff45\" kind=\"ReplicaSet\" apiVersion=\"apps/v1\" type=\"Normal\" reason=\"SuccessfulCreate\" message=\"Created pod: webserver-789dfdff45-cm78f\"\nI0718 20:02:43.981949       1 replica_set.go:559] \"Too few replicas\" replicaSet=\"deployment-4244/webserver-789dfdff45\" need=3 creating=1\nI0718 20:02:43.988887       1 event.go:291] \"Event occurred\" object=\"deployment-4244/webserver-789dfdff45\" kind=\"ReplicaSet\" apiVersion=\"apps/v1\" type=\"Normal\" reason=\"SuccessfulCreate\" message=\"Created pod: webserver-789dfdff45-hgmv4\"\nI0718 20:02:44.026521       1 replica_set.go:559] \"Too few replicas\" replicaSet=\"deployment-4244/webserver-847dcfb7fb\" need=2 creating=1\nI0718 20:02:44.034101       1 event.go:291] \"Event occurred\" object=\"deployment-4244/webserver-847dcfb7fb\" kind=\"ReplicaSet\" apiVersion=\"apps/v1\" type=\"Normal\" reason=\"SuccessfulCreate\" message=\"Created pod: webserver-847dcfb7fb-vbq95\"\nE0718 20:02:44.410526       1 tokens_controller.go:262] error synchronizing serviceaccount configmap-5641/default: secrets \"default-token-tm9bg\" is forbidden: unable to create new content in namespace configmap-5641 because it is being terminated\nE0718 20:02:44.504002       1 reflector.go:138] k8s.io/client-go/metadata/metadatainformer/informer.go:90: Failed to watch *v1.PartialObjectMetadata: failed to list *v1.PartialObjectMetadata: the server could not find the requested resource\nE0718 20:02:45.107407       1 namespace_controller.go:162] deletion of namespace cronjob-4272 failed: unexpected items still remain in namespace: cronjob-4272 for gvr: /v1, Resource=pods\nI0718 20:02:45.230873       1 garbagecollector.go:471] \"Processing object\" object=\"ephemeral-599-8917/csi-hostpath-attacher-s7k5l\" objectUID=dcdc7abe-112b-468c-9183-7592bb610c60 kind=\"EndpointSlice\" virtual=false\nI0718 20:02:45.234880       1 garbagecollector.go:580] \"Deleting object\" object=\"ephemeral-599-8917/csi-hostpath-attacher-s7k5l\" objectUID=dcdc7abe-112b-468c-9183-7592bb610c60 kind=\"EndpointSlice\" propagationPolicy=Background\nE0718 20:02:45.270417       1 namespace_controller.go:162] deletion of namespace cronjob-4272 failed: unexpected items still remain in namespace: cronjob-4272 for gvr: /v1, Resource=pods\nI0718 20:02:45.274374       1 garbagecollector.go:471] \"Processing object\" object=\"ephemeral-599-8917/csi-hostpath-attacher-56cf6f4b89\" objectUID=36381206-ed25-4c6e-9804-b11d59247126 kind=\"ControllerRevision\" virtual=false\nI0718 20:02:45.274583       1 stateful_set.go:419] StatefulSet has been deleted ephemeral-599-8917/csi-hostpath-attacher\nI0718 20:02:45.274618       1 garbagecollector.go:471] \"Processing object\" object=\"ephemeral-599-8917/csi-hostpath-attacher-0\" objectUID=207d52ee-35f7-431a-9308-5ba599426f64 kind=\"Pod\" virtual=false\nI0718 20:02:45.282236       1 garbagecollector.go:580] \"Deleting object\" object=\"ephemeral-599-8917/csi-hostpath-attacher-56cf6f4b89\" objectUID=36381206-ed25-4c6e-9804-b11d59247126 kind=\"ControllerRevision\" propagationPolicy=Background\nI0718 20:02:45.283178       1 garbagecollector.go:580] \"Deleting object\" object=\"ephemeral-599-8917/csi-hostpath-attacher-0\" objectUID=207d52ee-35f7-431a-9308-5ba599426f64 kind=\"Pod\" propagationPolicy=Background\nE0718 20:02:45.289324       1 reflector.go:138] k8s.io/client-go/metadata/metadatainformer/informer.go:90: Failed to watch *v1.PartialObjectMetadata: failed to list *v1.PartialObjectMetadata: the server could not find the requested resource\nI0718 20:02:45.341522       1 garbagecollector.go:471] \"Processing object\" object=\"ephemeral-599-8917/csi-hostpathplugin-x4p89\" objectUID=6808a9c9-9849-4a59-b04c-3e61029ab69f kind=\"EndpointSlice\" virtual=false\nI0718 20:02:45.346050       1 garbagecollector.go:580] \"Deleting object\" object=\"ephemeral-599-8917/csi-hostpathplugin-x4p89\" objectUID=6808a9c9-9849-4a59-b04c-3e61029ab69f kind=\"EndpointSlice\" propagationPolicy=Background\nI0718 20:02:45.386677       1 garbagecollector.go:471] \"Processing object\" object=\"ephemeral-599-8917/csi-hostpathplugin-76f9d6c8dc\" objectUID=801963cb-ddeb-4be2-b342-6b636399370b kind=\"ControllerRevision\" virtual=false\nI0718 20:02:45.386891       1 stateful_set.go:419] StatefulSet has been deleted ephemeral-599-8917/csi-hostpathplugin\nI0718 20:02:45.386936       1 garbagecollector.go:471] \"Processing object\" object=\"ephemeral-599-8917/csi-hostpathplugin-0\" objectUID=5fe0c0f8-3bc0-4f96-9afd-378223746b22 kind=\"Pod\" virtual=false\nI0718 20:02:45.388959       1 garbagecollector.go:580] \"Deleting object\" object=\"ephemeral-599-8917/csi-hostpathplugin-76f9d6c8dc\" objectUID=801963cb-ddeb-4be2-b342-6b636399370b kind=\"ControllerRevision\" propagationPolicy=Background\nI0718 20:02:45.389951       1 garbagecollector.go:580] \"Deleting object\" object=\"ephemeral-599-8917/csi-hostpathplugin-0\" objectUID=5fe0c0f8-3bc0-4f96-9afd-378223746b22 kind=\"Pod\" propagationPolicy=Background\nI0718 20:02:45.419468       1 garbagecollector.go:471] \"Processing object\" object=\"ephemeral-599-8917/csi-hostpath-provisioner-9p5s9\" objectUID=218a3ec0-ac5c-4065-a3eb-72a4e3df0190 kind=\"EndpointSlice\" virtual=false\nI0718 20:02:45.423407       1 garbagecollector.go:580] \"Deleting object\" object=\"ephemeral-599-8917/csi-hostpath-provisioner-9p5s9\" objectUID=218a3ec0-ac5c-4065-a3eb-72a4e3df0190 kind=\"EndpointSlice\" propagationPolicy=Background\nI0718 20:02:45.463906       1 garbagecollector.go:471] \"Processing object\" object=\"ephemeral-599-8917/csi-hostpath-provisioner-755f45bd9f\" objectUID=51752cdd-3434-4e4a-b2b6-7aec608b2cd5 kind=\"ControllerRevision\" virtual=false\nI0718 20:02:45.463931       1 garbagecollector.go:471] \"Processing object\" object=\"ephemeral-599-8917/csi-hostpath-provisioner-0\" objectUID=a34239e6-31f4-440d-9b58-c80e5121a2d0 kind=\"Pod\" virtual=false\nI0718 20:02:45.463942       1 stateful_set.go:419] StatefulSet has been deleted ephemeral-599-8917/csi-hostpath-provisioner\nI0718 20:02:45.469106       1 garbagecollector.go:580] \"Deleting object\" object=\"ephemeral-599-8917/csi-hostpath-provisioner-755f45bd9f\" objectUID=51752cdd-3434-4e4a-b2b6-7aec608b2cd5 kind=\"ControllerRevision\" propagationPolicy=Background\nI0718 20:02:45.469244       1 garbagecollector.go:580] \"Deleting object\" object=\"ephemeral-599-8917/csi-hostpath-provisioner-0\" objectUID=a34239e6-31f4-440d-9b58-c80e5121a2d0 kind=\"Pod\" propagationPolicy=Background\nE0718 20:02:45.489591       1 namespace_controller.go:162] deletion of namespace cronjob-4272 failed: unexpected items still remain in namespace: cronjob-4272 for gvr: /v1, Resource=pods\nI0718 20:02:45.496199       1 garbagecollector.go:471] \"Processing object\" object=\"ephemeral-599-8917/csi-hostpath-resizer-pqwpg\" objectUID=0313f34f-c731-41b6-9df5-535539ce1727 kind=\"EndpointSlice\" virtual=false\nI0718 20:02:45.504982       1 garbagecollector.go:580] \"Deleting object\" object=\"ephemeral-599-8917/csi-hostpath-resizer-pqwpg\" objectUID=0313f34f-c731-41b6-9df5-535539ce1727 kind=\"EndpointSlice\" propagationPolicy=Background\nI0718 20:02:45.545009       1 garbagecollector.go:471] \"Processing object\" object=\"ephemeral-599-8917/csi-hostpath-resizer-76fcdcf4b8\" objectUID=6a31ef33-add0-476c-9783-f3717fe51c59 kind=\"ControllerRevision\" virtual=false\nI0718 20:02:45.545055       1 stateful_set.go:419] StatefulSet has been deleted ephemeral-599-8917/csi-hostpath-resizer\nI0718 20:02:45.545080       1 garbagecollector.go:471] \"Processing object\" object=\"ephemeral-599-8917/csi-hostpath-resizer-0\" objectUID=a9b402d9-6945-4294-a3e2-d68dcca1c2c3 kind=\"Pod\" virtual=false\nI0718 20:02:45.547384       1 garbagecollector.go:580] \"Deleting object\" object=\"ephemeral-599-8917/csi-hostpath-resizer-76fcdcf4b8\" objectUID=6a31ef33-add0-476c-9783-f3717fe51c59 kind=\"ControllerRevision\" propagationPolicy=Background\nI0718 20:02:45.548388       1 garbagecollector.go:580] \"Deleting object\" object=\"ephemeral-599-8917/csi-hostpath-resizer-0\" objectUID=a9b402d9-6945-4294-a3e2-d68dcca1c2c3 kind=\"Pod\" propagationPolicy=Background\nI0718 20:02:45.549384       1 namespace_controller.go:185] Namespace has been deleted provisioning-7043\nI0718 20:02:45.575905       1 garbagecollector.go:471] \"Processing object\" object=\"ephemeral-599-8917/csi-hostpath-snapshotter-8ft55\" objectUID=faeee67d-d45c-42a3-8530-36ebb1ba2707 kind=\"EndpointSlice\" virtual=false\nI0718 20:02:45.579695       1 garbagecollector.go:580] \"Deleting object\" object=\"ephemeral-599-8917/csi-hostpath-snapshotter-8ft55\" objectUID=faeee67d-d45c-42a3-8530-36ebb1ba2707 kind=\"EndpointSlice\" propagationPolicy=Background\nI0718 20:02:45.621543       1 garbagecollector.go:471] \"Processing object\" object=\"ephemeral-599-8917/csi-hostpath-snapshotter-699884d77b\" objectUID=baef58d3-a62f-4e75-9146-b6bc1c383c4e kind=\"ControllerRevision\" virtual=false\nI0718 20:02:45.621650       1 garbagecollector.go:471] \"Processing object\" object=\"ephemeral-599-8917/csi-hostpath-snapshotter-0\" objectUID=06f5168a-7175-47b2-a9c8-72f784768da9 kind=\"Pod\" virtual=false\nI0718 20:02:45.621604       1 stateful_set.go:419] StatefulSet has been deleted ephemeral-599-8917/csi-hostpath-snapshotter\nI0718 20:02:45.623775       1 garbagecollector.go:580] \"Deleting object\" object=\"ephemeral-599-8917/csi-hostpath-snapshotter-0\" objectUID=06f5168a-7175-47b2-a9c8-72f784768da9 kind=\"Pod\" propagationPolicy=Background\nI0718 20:02:45.624640       1 garbagecollector.go:580] \"Deleting object\" object=\"ephemeral-599-8917/csi-hostpath-snapshotter-699884d77b\" objectUID=baef58d3-a62f-4e75-9146-b6bc1c383c4e kind=\"ControllerRevision\" propagationPolicy=Background\nE0718 20:02:45.718479       1 namespace_controller.go:162] deletion of namespace cronjob-4272 failed: unexpected items still remain in namespace: cronjob-4272 for gvr: /v1, Resource=pods\nI0718 20:02:45.799430       1 namespace_controller.go:185] Namespace has been deleted secrets-7800\nI0718 20:02:45.838814       1 event.go:291] \"Event occurred\" object=\"deployment-4244/webserver\" kind=\"Deployment\" apiVersion=\"apps/v1\" type=\"Normal\" reason=\"DeploymentRollback\" message=\"Rolled back deployment \\\"webserver\\\" to revision 2\"\nI0718 20:02:45.851035       1 deployment_controller.go:490] \"Error syncing deployment\" deployment=\"deployment-4244/webserver\" err=\"Operation cannot be fulfilled on replicasets.apps \\\"webserver-6584b976d5\\\": the object has been modified; please apply your changes to the latest version and try again\"\nI0718 20:02:45.864557       1 replica_set.go:595] \"Too many replicas\" replicaSet=\"deployment-4244/webserver-847dcfb7fb\" need=1 deleting=1\nI0718 20:02:45.864581       1 replica_set.go:223] \"Found related ReplicaSets\" replicaSet=\"deployment-4244/webserver-847dcfb7fb\" relatedReplicaSets=[webserver-847dcfb7fb webserver-6584b976d5 webserver-789dfdff45]\nI0718 20:02:45.864663       1 event.go:291] \"Event occurred\" object=\"deployment-4244/webserver\" kind=\"Deployment\" apiVersion=\"apps/v1\" type=\"Normal\" reason=\"ScalingReplicaSet\" message=\"Scaled down replica set webserver-847dcfb7fb to 1\"\nI0718 20:02:45.864763       1 controller_utils.go:602] \"Deleting pod\" controller=\"webserver-847dcfb7fb\" pod=\"deployment-4244/webserver-847dcfb7fb-vbq95\"\nI0718 20:02:45.868274       1 deployment_controller.go:490] \"Error syncing deployment\" deployment=\"deployment-4244/webserver\" err=\"Operation cannot be fulfilled on deployments.apps \\\"webserver\\\": the object has been modified; please apply your changes to the latest version and try again\"\nI0718 20:02:45.875584       1 replica_set.go:559] \"Too few replicas\" replicaSet=\"deployment-4244/webserver-6584b976d5\" need=3 creating=1\nI0718 20:02:45.875697       1 event.go:291] \"Event occurred\" object=\"deployment-4244/webserver\" kind=\"Deployment\" apiVersion=\"apps/v1\" type=\"Normal\" reason=\"ScalingReplicaSet\" message=\"Scaled up replica set webserver-6584b976d5 to 3\"\nI0718 20:02:45.877285       1 event.go:291] \"Event occurred\" object=\"deployment-4244/webserver-847dcfb7fb\" kind=\"ReplicaSet\" apiVersion=\"apps/v1\" type=\"Normal\" reason=\"SuccessfulDelete\" message=\"Deleted pod: webserver-847dcfb7fb-vbq95\"\nI0718 20:02:45.882544       1 event.go:291] \"Event occurred\" object=\"deployment-4244/webserver-6584b976d5\" kind=\"ReplicaSet\" apiVersion=\"apps/v1\" type=\"Normal\" reason=\"SuccessfulCreate\" message=\"Created pod: webserver-6584b976d5-f2m6v\"\nI0718 20:02:45.905751       1 event.go:291] \"Event occurred\" object=\"csi-mock-volumes-9933/pvc-hm6bq\" kind=\"PersistentVolumeClaim\" apiVersion=\"v1\" type=\"Normal\" reason=\"ExternalProvisioning\" message=\"waiting for a volume to be created, either by external provisioner \\\"csi-mock-csi-mock-volumes-9933\\\" or manually created by system administrator\"\nI0718 20:02:45.905772       1 event.go:291] \"Event occurred\" object=\"csi-mock-volumes-9933/pvc-hm6bq\" kind=\"PersistentVolumeClaim\" apiVersion=\"v1\" type=\"Normal\" reason=\"ExternalProvisioning\" message=\"waiting for a volume to be created, either by external provisioner \\\"csi-mock-csi-mock-volumes-9933\\\" or manually created by system administrator\"\nI0718 20:02:45.923833       1 pv_controller.go:879] volume \"pvc-dd7fe695-934e-4f09-8209-37cfbf1678ed\" entered phase \"Bound\"\nI0718 20:02:45.923870       1 pv_controller.go:982] volume \"pvc-dd7fe695-934e-4f09-8209-37cfbf1678ed\" bound to claim \"csi-mock-volumes-9933/pvc-hm6bq\"\nI0718 20:02:45.935185       1 pv_controller.go:823] claim \"csi-mock-volumes-9933/pvc-hm6bq\" entered phase \"Bound\"\nE0718 20:02:45.989835       1 namespace_controller.go:162] deletion of namespace cronjob-4272 failed: unexpected items still remain in namespace: cronjob-4272 for gvr: /v1, Resource=pods\nI0718 20:02:46.143545       1 namespace_controller.go:185] Namespace has been deleted projected-7562\nE0718 20:02:46.245906       1 namespace_controller.go:162] deletion of namespace cronjob-4272 failed: unexpected items still remain in namespace: cronjob-4272 for gvr: /v1, Resource=pods\nE0718 20:02:46.580219       1 namespace_controller.go:162] deletion of namespace cronjob-4272 failed: unexpected items still remain in namespace: cronjob-4272 for gvr: /v1, Resource=pods\nE0718 20:02:46.671059       1 namespace_controller.go:162] deletion of namespace kubectl-9687 failed: unexpected items still remain in namespace: kubectl-9687 for gvr: /v1, Resource=pods\nE0718 20:02:47.065101       1 namespace_controller.go:162] deletion of namespace cronjob-4272 failed: unexpected items still remain in namespace: cronjob-4272 for gvr: /v1, Resource=pods\nI0718 20:02:47.437522       1 pvc_protection_controller.go:303] \"Pod uses PVC\" pod=\"persistent-local-volumes-test-8575/pod-b8049a35-4a49-461e-be86-66a2c98b8ce5\" PVC=\"persistent-local-volumes-test-8575/pvc-p4ps5\"\nI0718 20:02:47.437553       1 pvc_protection_controller.go:181] \"Keeping PVC because it is being used\" PVC=\"persistent-local-volumes-test-8575/pvc-p4ps5\"\nE0718 20:02:47.872201       1 namespace_controller.go:162] deletion of namespace cronjob-4272 failed: unexpected items still remain in namespace: cronjob-4272 for gvr: /v1, Resource=pods\nE0718 20:02:47.942027       1 tokens_controller.go:262] error synchronizing serviceaccount persistent-local-volumes-test-7688/default: secrets \"default-token-2rm7z\" is forbidden: unable to create new content in namespace persistent-local-volumes-test-7688 because it is being terminated\nI0718 20:02:47.984071       1 namespace_controller.go:185] Namespace has been deleted volume-6255\nI0718 20:02:48.033124       1 namespace_controller.go:185] Namespace has been deleted ephemeral-599\nI0718 20:02:48.142829       1 replica_set.go:595] \"Too many replicas\" replicaSet=\"deployment-4244/webserver-789dfdff45\" need=2 deleting=1\nI0718 20:02:48.142860       1 replica_set.go:223] \"Found related ReplicaSets\" replicaSet=\"deployment-4244/webserver-789dfdff45\" relatedReplicaSets=[webserver-847dcfb7fb webserver-6584b976d5 webserver-789dfdff45]\nI0718 20:02:48.142917       1 controller_utils.go:602] \"Deleting pod\" controller=\"webserver-789dfdff45\" pod=\"deployment-4244/webserver-789dfdff45-cm78f\"\nI0718 20:02:48.142956       1 event.go:291] \"Event occurred\" object=\"deployment-4244/webserver\" kind=\"Deployment\" apiVersion=\"apps/v1\" type=\"Normal\" reason=\"ScalingReplicaSet\" message=\"Scaled down replica set webserver-789dfdff45 to 2\"\nI0718 20:02:48.155820       1 replica_set.go:559] \"Too few replicas\" replicaSet=\"deployment-4244/webserver-6584b976d5\" need=4 creating=1\nI0718 20:02:48.155893       1 event.go:291] \"Event occurred\" object=\"deployment-4244/webserver\" kind=\"Deployment\" apiVersion=\"apps/v1\" type=\"Normal\" reason=\"ScalingReplicaSet\" message=\"Scaled up replica set webserver-6584b976d5 to 4\"\nI0718 20:02:48.158089       1 event.go:291] \"Event occurred\" object=\"deployment-4244/webserver-789dfdff45\" kind=\"ReplicaSet\" apiVersion=\"apps/v1\" type=\"Normal\" reason=\"SuccessfulDelete\" message=\"Deleted pod: webserver-789dfdff45-cm78f\"\nI0718 20:02:48.159983       1 reconciler.go:295] attacherDetacher.AttachVolume started for volume \"pvc-dd7fe695-934e-4f09-8209-37cfbf1678ed\" (UniqueName: \"kubernetes.io/csi/csi-mock-csi-mock-volumes-9933^4\") from node \"10.116.0.7\" \nI0718 20:02:48.166207       1 event.go:291] \"Event occurred\" object=\"deployment-4244/webserver-6584b976d5\" kind=\"ReplicaSet\" apiVersion=\"apps/v1\" type=\"Normal\" reason=\"SuccessfulCreate\" message=\"Created pod: webserver-6584b976d5-cxgwl\"\nI0718 20:02:48.452204       1 resource_quota_controller.go:307] Resource quota has been deleted resourcequota-5091/test-quota\nI0718 20:02:48.703217       1 operation_generator.go:368] AttachVolume.Attach succeeded for volume \"pvc-dd7fe695-934e-4f09-8209-37cfbf1678ed\" (UniqueName: \"kubernetes.io/csi/csi-mock-csi-mock-volumes-9933^4\") from node \"10.116.0.7\" \nI0718 20:02:48.703373       1 event.go:291] \"Event occurred\" object=\"csi-mock-volumes-9933/pvc-volume-tester-rfqs7\" kind=\"Pod\" apiVersion=\"v1\" type=\"Normal\" reason=\"SuccessfulAttachVolume\" message=\"AttachVolume.Attach succeeded for volume \\\"pvc-dd7fe695-934e-4f09-8209-37cfbf1678ed\\\" \"\nI0718 20:02:49.228492       1 event.go:291] \"Event occurred\" object=\"deployment-4244/webserver\" kind=\"Deployment\" apiVersion=\"apps/v1\" type=\"Normal\" reason=\"DeploymentRollback\" message=\"Rolled back deployment \\\"webserver\\\" to revision 3\"\nI0718 20:02:49.240000       1 deployment_controller.go:490] \"Error syncing deployment\" deployment=\"deployment-4244/webserver\" err=\"Operation cannot be fulfilled on replicasets.apps \\\"webserver-789dfdff45\\\": the object has been modified; please apply your changes to the latest version and try again\"\nI0718 20:02:49.285934       1 replica_set.go:595] \"Too many replicas\" replicaSet=\"deployment-4244/webserver-6584b976d5\" need=3 deleting=1\nI0718 20:02:49.285966       1 replica_set.go:223] \"Found related ReplicaSets\" replicaSet=\"deployment-4244/webserver-6584b976d5\" relatedReplicaSets=[webserver-847dcfb7fb webserver-6584b976d5 webserver-789dfdff45]\nI0718 20:02:49.286010       1 event.go:291] \"Event occurred\" object=\"deployment-4244/webserver\" kind=\"Deployment\" apiVersion=\"apps/v1\" type=\"Normal\" reason=\"ScalingReplicaSet\" message=\"Scaled down replica set webserver-6584b976d5 to 3\"\nI0718 20:02:49.286029       1 controller_utils.go:602] \"Deleting pod\" controller=\"webserver-6584b976d5\" pod=\"deployment-4244/webserver-6584b976d5-cxgwl\"\nI0718 20:02:49.290114       1 deployment_controller.go:490] \"Error syncing deployment\" deployment=\"deployment-4244/webserver\" err=\"Operation cannot be fulfilled on deployments.apps \\\"webserver\\\": the object has been modified; please apply your changes to the latest version and try again\"\nI0718 20:02:49.299632       1 replica_set.go:559] \"Too few replicas\" replicaSet=\"deployment-4244/webserver-789dfdff45\" need=3 creating=1\nI0718 20:02:49.299688       1 event.go:291] \"Event occurred\" object=\"deployment-4244/webserver\" kind=\"Deployment\" apiVersion=\"apps/v1\" type=\"Normal\" reason=\"ScalingReplicaSet\" message=\"Scaled up replica set webserver-789dfdff45 to 3\"\nI0718 20:02:49.301929       1 event.go:291] \"Event occurred\" object=\"deployment-4244/webserver-6584b976d5\" kind=\"ReplicaSet\" apiVersion=\"apps/v1\" type=\"Normal\" reason=\"SuccessfulDelete\" message=\"Deleted pod: webserver-6584b976d5-cxgwl\"\nI0718 20:02:49.305711       1 event.go:291] \"Event occurred\" object=\"deployment-4244/webserver-789dfdff45\" kind=\"ReplicaSet\" apiVersion=\"apps/v1\" type=\"Normal\" reason=\"SuccessfulCreate\" message=\"Created pod: webserver-789dfdff45-dfzxn\"\nE0718 20:02:49.325941       1 tokens_controller.go:262] error synchronizing serviceaccount persistent-local-volumes-test-8656/default: secrets \"default-token-lnsk9\" is forbidden: unable to create new content in namespace persistent-local-volumes-test-8656 because it is being terminated\nI0718 20:02:49.347319       1 pvc_protection_controller.go:291] \"PVC is unused\" PVC=\"persistent-local-volumes-test-8656/pvc-q52dg\"\nI0718 20:02:49.355931       1 pv_controller.go:640] volume \"local-pvnlx44\" is released and reclaim policy \"Retain\" will be executed\nI0718 20:02:49.361361       1 pv_controller.go:879] volume \"local-pvnlx44\" entered phase \"Released\"\nI0718 20:02:49.368127       1 pv_controller_base.go:505] deletion of claim \"persistent-local-volumes-test-8656/pvc-q52dg\" was already processed\nE0718 20:02:49.399191       1 namespace_controller.go:162] deletion of namespace cronjob-4272 failed: unexpected items still remain in namespace: cronjob-4272 for gvr: /v1, Resource=pods\nI0718 20:02:51.371614       1 replica_set.go:595] \"Too many replicas\" replicaSet=\"deployment-4244/webserver-6584b976d5\" need=2 deleting=1\nI0718 20:02:51.371646       1 replica_set.go:223] \"Found related ReplicaSets\" replicaSet=\"deployment-4244/webserver-6584b976d5\" relatedReplicaSets=[webserver-847dcfb7fb webserver-6584b976d5 webserver-789dfdff45]\nI0718 20:02:51.371698       1 event.go:291] \"Event occurred\" object=\"deployment-4244/webserver\" kind=\"Deployment\" apiVersion=\"apps/v1\" type=\"Normal\" reason=\"ScalingReplicaSet\" message=\"Scaled down replica set webserver-6584b976d5 to 2\"\nI0718 20:02:51.371716       1 controller_utils.go:602] \"Deleting pod\" controller=\"webserver-6584b976d5\" pod=\"deployment-4244/webserver-6584b976d5-f2m6v\"\nI0718 20:02:51.378657       1 replica_set.go:595] \"Too many replicas\" replicaSet=\"deployment-4244/webserver-789dfdff45\" need=2 deleting=1\nI0718 20:02:51.378686       1 replica_set.go:223] \"Found related ReplicaSets\" replicaSet=\"deployment-4244/webserver-789dfdff45\" relatedReplicaSets=[webserver-847dcfb7fb webserver-6584b976d5 webserver-789dfdff45]\nI0718 20:02:51.378774       1 controller_utils.go:602] \"Deleting pod\" controller=\"webserver-789dfdff45\" pod=\"deployment-4244/webserver-789dfdff45-dfzxn\"\nI0718 20:02:51.378865       1 event.go:291] \"Event occurred\" object=\"deployment-4244/webserver\" kind=\"Deployment\" apiVersion=\"apps/v1\" type=\"Normal\" reason=\"ScalingReplicaSet\" message=\"Scaled down replica set webserver-789dfdff45 to 2\"\nI0718 20:02:51.388893       1 event.go:291] \"Event occurred\" object=\"deployment-4244/webserver-6584b976d5\" kind=\"ReplicaSet\" apiVersion=\"apps/v1\" type=\"Normal\" reason=\"SuccessfulDelete\" message=\"Deleted pod: webserver-6584b976d5-f2m6v\"\nI0718 20:02:51.405133       1 event.go:291] \"Event occurred\" object=\"deployment-4244/webserver-789dfdff45\" kind=\"ReplicaSet\" apiVersion=\"apps/v1\" type=\"Normal\" reason=\"SuccessfulDelete\" message=\"Deleted pod: webserver-789dfdff45-dfzxn\"\nI0718 20:02:51.415853       1 deployment_controller.go:490] \"Error syncing deployment\" deployment=\"deployment-4244/webserver\" err=\"Operation cannot be fulfilled on deployments.apps \\\"webserver\\\": the object has been modified; please apply your changes to the latest version and try again\"\nI0718 20:02:51.436063       1 replica_set.go:595] \"Too many replicas\" replicaSet=\"deployment-4244/webserver-6584b976d5\" need=1 deleting=1\nI0718 20:02:51.436091       1 replica_set.go:223] \"Found related ReplicaSets\" replicaSet=\"deployment-4244/webserver-6584b976d5\" relatedReplicaSets=[webserver-847dcfb7fb webserver-6584b976d5 webserver-789dfdff45 webserver-555c7ddb77]\nI0718 20:02:51.436149       1 controller_utils.go:602] \"Deleting pod\" controller=\"webserver-6584b976d5\" pod=\"deployment-4244/webserver-6584b976d5-52nlw\"\nI0718 20:02:51.436222       1 event.go:291] \"Event occurred\" object=\"deployment-4244/webserver\" kind=\"Deployment\" apiVersion=\"apps/v1\" type=\"Normal\" reason=\"ScalingReplicaSet\" message=\"Scaled down replica set webserver-6584b976d5 to 1\"\nI0718 20:02:51.444242       1 deployment_controller.go:490] \"Error syncing deployment\" deployment=\"deployment-4244/webserver\" err=\"Operation cannot be fulfilled on deployments.apps \\\"webserver\\\": the object has been modified; please apply your changes to the latest version and try again\"\nI0718 20:02:51.450955       1 replica_set.go:559] \"Too few replicas\" replicaSet=\"deployment-4244/webserver-555c7ddb77\" need=1 creating=1\nI0718 20:02:51.451157       1 event.go:291] \"Event occurred\" object=\"deployment-4244/webserver\" kind=\"Deployment\" apiVersion=\"apps/v1\" type=\"Normal\" reason=\"ScalingReplicaSet\" message=\"Scaled up replica set webserver-555c7ddb77 to 1\"\nI0718 20:02:51.453330       1 event.go:291] \"Event occurred\" object=\"deployment-4244/webserver-6584b976d5\" kind=\"ReplicaSet\" apiVersion=\"apps/v1\" type=\"Normal\" reason=\"SuccessfulDelete\" message=\"Deleted pod: webserver-6584b976d5-52nlw\"\nI0718 20:02:51.458101       1 event.go:291] \"Event occurred\" object=\"deployment-4244/webserver-555c7ddb77\" kind=\"ReplicaSet\" apiVersion=\"apps/v1\" type=\"Normal\" reason=\"SuccessfulCreate\" message=\"Created pod: webserver-555c7ddb77-ld8hl\"\nI0718 20:02:51.472340       1 replica_set.go:595] \"Too many replicas\" replicaSet=\"deployment-4244/webserver-789dfdff45\" need=1 deleting=1\nI0718 20:02:51.472371       1 replica_set.go:223] \"Found related ReplicaSets\" replicaSet=\"deployment-4244/webserver-789dfdff45\" relatedReplicaSets=[webserver-847dcfb7fb webserver-6584b976d5 webserver-789dfdff45 webserver-555c7ddb77]\nI0718 20:02:51.472451       1 controller_utils.go:602] \"Deleting pod\" controller=\"webserver-789dfdff45\" pod=\"deployment-4244/webserver-789dfdff45-hgmv4\"\nI0718 20:02:51.472478       1 event.go:291] \"Event occurred\" object=\"deployment-4244/webserver\" kind=\"Deployment\" apiVersion=\"apps/v1\" type=\"Normal\" reason=\"ScalingReplicaSet\" message=\"Scaled down replica set webserver-789dfdff45 to 1\"\nI0718 20:02:51.476672       1 deployment_controller.go:490] \"Error syncing deployment\" deployment=\"deployment-4244/webserver\" err=\"Operation cannot be fulfilled on deployments.apps \\\"webserver\\\": the object has been modified; please apply your changes to the latest version and try again\"\nI0718 20:02:51.570003       1 pv_controller.go:930] claim \"provisioning-9406/pvc-c4v95\" bound to volume \"local-5jzbn\"\nI0718 20:02:51.736688       1 deployment_controller.go:490] \"Error syncing deployment\" deployment=\"deployment-4244/webserver\" err=\"Operation cannot be fulfilled on replicasets.apps \\\"webserver-555c7ddb77\\\": the object has been modified; please apply your changes to the latest version and try again\"\nI0718 20:02:51.752143       1 replica_set.go:559] \"Too few replicas\" replicaSet=\"deployment-4244/webserver-555c7ddb77\" need=2 creating=1\nI0718 20:02:51.752797       1 event.go:291] \"Event occurred\" object=\"deployment-4244/webserver-789dfdff45\" kind=\"ReplicaSet\" apiVersion=\"apps/v1\" type=\"Normal\" reason=\"SuccessfulDelete\" message=\"Deleted pod: webserver-789dfdff45-hgmv4\"\nI0718 20:02:51.752807       1 event.go:291] \"Event occurred\" object=\"deployment-4244/webserver\" kind=\"Deployment\" apiVersion=\"apps/v1\" type=\"Normal\" reason=\"ScalingReplicaSet\" message=\"Scaled up replica set webserver-555c7ddb77 to 2\"\nI0718 20:02:51.756860       1 pv_controller.go:879] volume \"local-5jzbn\" entered phase \"Bound\"\nI0718 20:02:51.756918       1 pv_controller.go:982] volume \"local-5jzbn\" bound to claim \"provisioning-9406/pvc-c4v95\"\nI0718 20:02:51.765288       1 event.go:291] \"Event occurred\" object=\"deployment-4244/webserver-555c7ddb77\" kind=\"ReplicaSet\" apiVersion=\"apps/v1\" type=\"Normal\" reason=\"SuccessfulCreate\" message=\"Created pod: webserver-555c7ddb77-l85dz\"\nI0718 20:02:51.804758       1 pv_controller.go:823] claim \"provisioning-9406/pvc-c4v95\" entered phase \"Bound\"\nI0718 20:02:52.001302       1 namespace_controller.go:185] Namespace has been deleted provisioning-3434\nI0718 20:02:52.375162       1 resource_quota_controller.go:435] syncing resource quota controller with updated resources from discovery: added: [crd-publish-openapi-test-multi-ver.example.com/v4, Resource=e2e-test-crd-publish-openapi-9331-crds], removed: [kubectl.example.com/v1, Resource=e2e-test-kubectl-4850-crds]\nI0718 20:02:52.375260       1 resource_quota_monitor.go:229] QuotaMonitor created object count evaluator for e2e-test-crd-publish-openapi-9331-crds.crd-publish-openapi-test-multi-ver.example.com\nI0718 20:02:52.375305       1 shared_informer.go:240] Waiting for caches to sync for resource quota\nI0718 20:02:52.476346       1 shared_informer.go:247] Caches are synced for resource quota \nI0718 20:02:52.476377       1 resource_quota_controller.go:454] synced quota controller\nI0718 20:02:52.760060       1 namespace_controller.go:185] Namespace has been deleted kubectl-3942\nI0718 20:02:52.850332       1 garbagecollector.go:213] syncing garbage collector with updated resources from discovery (attempt 1): added: [crd-publish-openapi-test-multi-ver.example.com/v4, Resource=e2e-test-crd-publish-openapi-9331-crds], removed: []\nI0718 20:02:52.856670       1 shared_informer.go:240] Waiting for caches to sync for garbage collector\nI0718 20:02:52.856721       1 shared_informer.go:247] Caches are synced for garbage collector \nI0718 20:02:52.856727       1 garbagecollector.go:254] synced garbage collector\nI0718 20:02:53.098335       1 namespace_controller.go:185] Namespace has been deleted persistent-local-volumes-test-7688\nE0718 20:02:53.391407       1 tokens_controller.go:262] error synchronizing serviceaccount persistent-local-volumes-test-8575/default: secrets \"default-token-k6sjt\" is forbidden: unable to create new content in namespace persistent-local-volumes-test-8575 because it is being terminated\nI0718 20:02:53.468375       1 namespace_controller.go:185] Namespace has been deleted resourcequota-5091\nI0718 20:02:53.616185       1 pvc_protection_controller.go:291] \"PVC is unused\" PVC=\"persistent-local-volumes-test-8575/pvc-p4ps5\"\nI0718 20:02:53.626733       1 pv_controller.go:640] volume \"local-pvkl66h\" is released and reclaim policy \"Retain\" will be executed\nI0718 20:02:53.632763       1 pv_controller.go:879] volume \"local-pvkl66h\" entered phase \"Released\"\nI0718 20:02:53.638805       1 pv_controller_base.go:505] deletion of claim \"persistent-local-volumes-test-8575/pvc-p4ps5\" was already processed\nI0718 20:02:53.769185       1 namespace_controller.go:185] Namespace has been deleted pod-network-test-1318\nE0718 20:02:53.934955       1 tokens_controller.go:262] error synchronizing serviceaccount subpath-8313/default: secrets \"default-token-tcn5r\" is forbidden: unable to create new content in namespace subpath-8313 because it is being terminated\nI0718 20:02:54.530970       1 namespace_controller.go:185] Namespace has been deleted persistent-local-volumes-test-8656\nI0718 20:02:54.884425       1 glusterfs.go:751] allocated GID 2000 for PVC pvc-54nzk\nI0718 20:02:54.901141       1 glusterfs.go:810] create volume of size 2GiB\nI0718 20:02:55.307845       1 glusterfs.go:841] volume with size 2 and name vol_78d2781814f7e985dfc5654b7d0a57c2 created\nI0718 20:02:55.337157       1 pv_controller.go:1677] volume \"pvc-1a55bfdc-a5ec-42bc-9bde-88062ff2b3d7\" provisioned for claim \"volume-provisioning-6451/pvc-54nzk\"\nI0718 20:02:55.337223       1 event.go:291] \"Event occurred\" object=\"volume-provisioning-6451/pvc-54nzk\" kind=\"PersistentVolumeClaim\" apiVersion=\"v1\" type=\"Normal\" reason=\"ProvisioningSucceeded\" message=\"Successfully provisioned volume pvc-1a55bfdc-a5ec-42bc-9bde-88062ff2b3d7 using kubernetes.io/glusterfs\"\nI0718 20:02:55.342295       1 pv_controller.go:879] volume \"pvc-1a55bfdc-a5ec-42bc-9bde-88062ff2b3d7\" entered phase \"Bound\"\nI0718 20:02:55.342353       1 pv_controller.go:982] volume \"pvc-1a55bfdc-a5ec-42bc-9bde-88062ff2b3d7\" bound to claim \"volume-provisioning-6451/pvc-54nzk\"\nI0718 20:02:55.354622       1 pv_controller.go:823] claim \"volume-provisioning-6451/pvc-54nzk\" entered phase \"Bound\"\nI0718 20:02:55.554864       1 event.go:291] \"Event occurred\" object=\"deployment-4244/webserver\" kind=\"Deployment\" apiVersion=\"apps/v1\" type=\"Normal\" reason=\"DeploymentRollback\" message=\"Rolled back deployment \\\"webserver\\\" to revision 5\"\nI0718 20:02:55.566182       1 deployment_controller.go:490] \"Error syncing deployment\" deployment=\"deployment-4244/webserver\" err=\"Operation cannot be fulfilled on replicasets.apps \\\"webserver-789dfdff45\\\": the object has been modified; please apply your changes to the latest version and try again\"\nI0718 20:02:55.579554       1 replica_set.go:595] \"Too many replicas\" replicaSet=\"deployment-4244/webserver-555c7ddb77\" need=1 deleting=1\nI0718 20:02:55.579583       1 replica_set.go:223] \"Found related ReplicaSets\" replicaSet=\"deployment-4244/webserver-555c7ddb77\" relatedReplicaSets=[webserver-847dcfb7fb webserver-6584b976d5 webserver-789dfdff45 webserver-555c7ddb77]\nI0718 20:02:55.579714       1 event.go:291] \"Event occurred\" object=\"deployment-4244/webserver\" kind=\"Deployment\" apiVersion=\"apps/v1\" type=\"Normal\" reason=\"ScalingReplicaSet\" message=\"Scaled down replica set webserver-555c7ddb77 to 1\"\nI0718 20:02:55.579744       1 controller_utils.go:602] \"Deleting pod\" controller=\"webserver-555c7ddb77\" pod=\"deployment-4244/webserver-555c7ddb77-ld8hl\"\nI0718 20:02:55.583507       1 deployment_controller.go:490] \"Error syncing deployment\" deployment=\"deployment-4244/webserver\" err=\"Operation cannot be fulfilled on deployments.apps \\\"webserver\\\": the object has been modified; please apply your changes to the latest version and try again\"\nI0718 20:02:55.592781       1 replica_set.go:559] \"Too few replicas\" replicaSet=\"deployment-4244/webserver-789dfdff45\" need=2 creating=1\nI0718 20:02:55.593029       1 event.go:291] \"Event occurred\" object=\"deployment-4244/webserver\" kind=\"Deployment\" apiVersion=\"apps/v1\" type=\"Normal\" reason=\"ScalingReplicaSet\" message=\"Scaled up replica set webserver-789dfdff45 to 2\"\nI0718 20:02:55.598800       1 event.go:291] \"Event occurred\" object=\"deployment-4244/webserver-789dfdff45\" kind=\"ReplicaSet\" apiVersion=\"apps/v1\" type=\"Normal\" reason=\"SuccessfulCreate\" message=\"Created pod: webserver-789dfdff45-l6dtq\"\nI0718 20:02:55.608009       1 event.go:291] \"Event occurred\" object=\"deployment-4244/webserver-555c7ddb77\" kind=\"ReplicaSet\" apiVersion=\"apps/v1\" type=\"Normal\" reason=\"SuccessfulDelete\" message=\"Deleted pod: webserver-555c7ddb77-ld8hl\"\nI0718 20:02:55.635907       1 deployment_controller.go:490] \"Error syncing deployment\" deployment=\"deployment-4244/webserver\" err=\"Operation cannot be fulfilled on deployments.apps \\\"webserver\\\": the object has been modified; please apply your changes to the latest version and try again\"\nI0718 20:02:55.757151       1 garbagecollector.go:471] \"Processing object\" object=\"services-4562/externalsvc-lmv6f\" objectUID=63d50080-ffa0-45f2-a11a-79fab514e444 kind=\"EndpointSlice\" virtual=false\nI0718 20:02:55.765521       1 garbagecollector.go:580] \"Deleting object\" object=\"services-4562/externalsvc-lmv6f\" objectUID=63d50080-ffa0-45f2-a11a-79fab514e444 kind=\"EndpointSlice\" propagationPolicy=Background\nI0718 20:02:55.808966       1 garbagecollector.go:471] \"Processing object\" object=\"services-4562/clusterip-service-ts8zm\" objectUID=f675fb20-e0d0-42f1-86fa-3cbf7371225d kind=\"EndpointSlice\" virtual=false\nI0718 20:02:55.809022       1 garbagecollector.go:471] \"Processing object\" object=\"services-4562/clusterip-service-w2bzn\" objectUID=87cd32b0-68f5-4fc8-adb9-4b0a25c7fc42 kind=\"EndpointSlice\" virtual=false\nI0718 20:02:55.811364       1 garbagecollector.go:580] \"Deleting object\" object=\"services-4562/clusterip-service-ts8zm\" objectUID=f675fb20-e0d0-42f1-86fa-3cbf7371225d kind=\"EndpointSlice\" propagationPolicy=Background\nI0718 20:02:55.813585       1 garbagecollector.go:580] \"Deleting object\" object=\"services-4562/clusterip-service-w2bzn\" objectUID=87cd32b0-68f5-4fc8-adb9-4b0a25c7fc42 kind=\"EndpointSlice\" propagationPolicy=Background\nE0718 20:02:56.191931       1 disruption.go:534] Error syncing PodDisruptionBudget disruption-6670/foo, requeuing: Operation cannot be fulfilled on poddisruptionbudgets.policy \"foo\": the object has been modified; please apply your changes to the latest version and try again\nE0718 20:02:56.810440       1 pv_controller.go:1452] error finding provisioning plugin for claim provisioning-6512/pvc-qjd42: storageclass.storage.k8s.io \"provisioning-6512\" not found\nI0718 20:02:56.810525       1 event.go:291] \"Event occurred\" object=\"provisioning-6512/pvc-qjd42\" kind=\"PersistentVolumeClaim\" apiVersion=\"v1\" type=\"Warning\" reason=\"ProvisioningFailed\" message=\"storageclass.storage.k8s.io \\\"provisioning-6512\\\" not found\"\nI0718 20:02:56.848907       1 pv_controller.go:879] volume \"local-r4c7r\" entered phase \"Available\"\nI0718 20:02:57.015755       1 namespace_controller.go:185] Namespace has been deleted kubectl-9687\nI0718 20:02:57.037972       1 pvc_protection_controller.go:291] \"PVC is unused\" PVC=\"volume-provisioning-6451/pvc-54nzk\"\nI0718 20:02:57.046879       1 pv_controller.go:640] volume \"pvc-1a55bfdc-a5ec-42bc-9bde-88062ff2b3d7\" is released and reclaim policy \"Delete\" will be executed\nI0718 20:02:57.053395       1 pv_controller.go:879] volume \"pvc-1a55bfdc-a5ec-42bc-9bde-88062ff2b3d7\" entered phase \"Released\"\nI0718 20:02:57.055802       1 pv_controller.go:1341] isVolumeReleased[pvc-1a55bfdc-a5ec-42bc-9bde-88062ff2b3d7]: volume is released\nI0718 20:02:57.055824       1 glusterfs.go:639] delete volume vol_78d2781814f7e985dfc5654b7d0a57c2\nI0718 20:02:57.142692       1 namespace_controller.go:185] Namespace has been deleted cronjob-4272\nI0718 20:02:57.465453       1 glusterfs.go:682] volume vol_78d2781814f7e985dfc5654b7d0a57c2 deleted successfully\nI0718 20:02:57.479144       1 garbagecollector.go:471] \"Processing object\" object=\"volume-provisioning-6451/glusterfs-dynamic-1a55bfdc-a5ec-42bc-9bde-88062ff2b3d7-ss92c\" objectUID=1c7c7fdf-0a89-4373-a98d-f36e0a219346 kind=\"EndpointSlice\" virtual=false\nI0718 20:02:57.483613       1 garbagecollector.go:580] \"Deleting object\" object=\"volume-provisioning-6451/glusterfs-dynamic-1a55bfdc-a5ec-42bc-9bde-88062ff2b3d7-ss92c\" objectUID=1c7c7fdf-0a89-4373-a98d-f36e0a219346 kind=\"EndpointSlice\" propagationPolicy=Background\nI0718 20:02:57.488337       1 glusterfs.go:949] service/endpoint: volume-provisioning-6451/glusterfs-dynamic-1a55bfdc-a5ec-42bc-9bde-88062ff2b3d7 deleted successfully\nI0718 20:02:57.488361       1 glusterfs.go:704] endpoint volume-provisioning-6451/glusterfs-dynamic-1a55bfdc-a5ec-42bc-9bde-88062ff2b3d7 is deleted successfully \nI0718 20:02:57.488390       1 pv_controller.go:1436] volume \"pvc-1a55bfdc-a5ec-42bc-9bde-88062ff2b3d7\" deleted\nE0718 20:02:57.488888       1 garbagecollector.go:350] error syncing item &garbagecollector.node{identity:garbagecollector.objectReference{OwnerReference:v1.OwnerReference{APIVersion:\"discovery.k8s.io/v1\", Kind:\"EndpointSlice\", Name:\"glusterfs-dynamic-1a55bfdc-a5ec-42bc-9bde-88062ff2b3d7-ss92c\", UID:\"1c7c7fdf-0a89-4373-a98d-f36e0a219346\", Controller:(*bool)(nil), BlockOwnerDeletion:(*bool)(nil)}, Namespace:\"volume-provisioning-6451\"}, dependentsLock:sync.RWMutex{w:sync.Mutex{state:0, sema:0x0}, writerSem:0x0, readerSem:0x0, readerCount:1, readerWait:0}, dependents:map[*garbagecollector.node]struct {}{}, deletingDependents:false, deletingDependentsLock:sync.RWMutex{w:sync.Mutex{state:0, sema:0x0}, writerSem:0x0, readerSem:0x0, readerCount:0, readerWait:0}, beingDeleted:false, beingDeletedLock:sync.RWMutex{w:sync.Mutex{state:0, sema:0x0}, writerSem:0x0, readerSem:0x0, readerCount:0, readerWait:0}, virtual:false, virtualLock:sync.RWMutex{w:sync.Mutex{state:0, sema:0x0}, writerSem:0x0, readerSem:0x0, readerCount:0, readerWait:0}, owners:[]v1.OwnerReference{v1.OwnerReference{APIVersion:\"v1\", Kind:\"Endpoints\", Name:\"glusterfs-dynamic-1a55bfdc-a5ec-42bc-9bde-88062ff2b3d7\", UID:\"705ef5a3-b11b-4675-b8d9-c07a4f6012b4\", Controller:(*bool)(0xc0032d976c), BlockOwnerDeletion:(*bool)(0xc0032d976d)}}}: endpointslices.discovery.k8s.io \"glusterfs-dynamic-1a55bfdc-a5ec-42bc-9bde-88062ff2b3d7-ss92c\" not found\nI0718 20:02:57.494419       1 garbagecollector.go:471] \"Processing object\" object=\"volume-provisioning-6451/glusterfs-dynamic-1a55bfdc-a5ec-42bc-9bde-88062ff2b3d7-ss92c\" objectUID=1c7c7fdf-0a89-4373-a98d-f36e0a219346 kind=\"EndpointSlice\" virtual=false\nI0718 20:02:57.502187       1 pv_controller_base.go:505] deletion of claim \"volume-provisioning-6451/pvc-54nzk\" was already processed\nI0718 20:02:57.697526       1 event.go:291] \"Event occurred\" object=\"deployment-4244/webserver\" kind=\"Deployment\" apiVersion=\"apps/v1\" type=\"Normal\" reason=\"DeploymentRollback\" message=\"Rolled back deployment \\\"webserver\\\" to revision 6\"\nI0718 20:02:57.708247       1 deployment_controller.go:490] \"Error syncing deployment\" deployment=\"deployment-4244/webserver\" err=\"Operation cannot be fulfilled on replicasets.apps \\\"webserver-555c7ddb77\\\": the object has been modified; please apply your changes to the latest version and try again\"\nI0718 20:02:57.725567       1 replica_set.go:595] \"Too many replicas\" replicaSet=\"deployment-4244/webserver-789dfdff45\" need=1 deleting=1\nI0718 20:02:57.725600       1 replica_set.go:223] \"Found related ReplicaSets\" replicaSet=\"deployment-4244/webserver-789dfdff45\" relatedReplicaSets=[webserver-847dcfb7fb webserver-6584b976d5 webserver-789dfdff45 webserver-555c7ddb77]\nI0718 20:02:57.725669       1 controller_utils.go:602] \"Deleting pod\" controller=\"webserver-789dfdff45\" pod=\"deployment-4244/webserver-789dfdff45-l6dtq\"\nI0718 20:02:57.725737       1 event.go:291] \"Event occurred\" object=\"deployment-4244/webserver\" kind=\"Deployment\" apiVersion=\"apps/v1\" type=\"Normal\" reason=\"ScalingReplicaSet\" message=\"Scaled down replica set webserver-789dfdff45 to 1\"\nI0718 20:02:57.732331       1 deployment_controller.go:490] \"Error syncing deployment\" deployment=\"deployment-4244/webserver\" err=\"Operation cannot be fulfilled on deployments.apps \\\"webserver\\\": the object has been modified; please apply your changes to the latest version and try again\"\nI0718 20:02:57.748197       1 replica_set.go:559] \"Too few replicas\" replicaSet=\"deployment-4244/webserver-555c7ddb77\" need=2 creating=1\nI0718 20:02:57.748379       1 event.go:291] \"Event occurred\" object=\"deployment-4244/webserver\" kind=\"Deployment\" apiVersion=\"apps/v1\" type=\"Normal\" reason=\"ScalingReplicaSet\" message=\"Scaled up replica set webserver-555c7ddb77 to 2\"\nI0718 20:02:57.750104       1 event.go:291] \"Event occurred\" object=\"deployment-4244/webserver-789dfdff45\" kind=\"ReplicaSet\" apiVersion=\"apps/v1\" type=\"Normal\" reason=\"SuccessfulDelete\" message=\"Deleted pod: webserver-789dfdff45-l6dtq\"\nI0718 20:02:57.763128       1 event.go:291] \"Event occurred\" object=\"deployment-4244/webserver-555c7ddb77\" kind=\"ReplicaSet\" apiVersion=\"apps/v1\" type=\"Normal\" reason=\"SuccessfulCreate\" message=\"Created pod: webserver-555c7ddb77-4g494\"\nI0718 20:02:57.843937       1 namespace_controller.go:185] Namespace has been deleted crd-publish-openapi-3018\nI0718 20:02:58.032433       1 deployment_controller.go:490] \"Error syncing deployment\" deployment=\"deployment-4244/webserver\" err=\"Operation cannot be fulfilled on deployments.apps \\\"webserver\\\": the object has been modified; please apply your changes to the latest version and try again\"\nI0718 20:02:58.647243       1 namespace_controller.go:185] Namespace has been deleted persistent-local-volumes-test-8575\nI0718 20:02:58.934703       1 pvc_protection_controller.go:303] \"Pod uses PVC\" pod=\"ephemeral-5827/inline-volume-tester-knktf\" PVC=\"ephemeral-5827/inline-volume-tester-knktf-my-volume-0\"\nI0718 20:02:58.934739       1 pvc_protection_controller.go:181] \"Keeping PVC because it is being used\" PVC=\"ephemeral-5827/inline-volume-tester-knktf-my-volume-0\"\nE0718 20:02:59.025857       1 reflector.go:138] k8s.io/client-go/metadata/metadatainformer/informer.go:90: Failed to watch *v1.PartialObjectMetadata: the server could not find the requested resource\nI0718 20:02:59.056650       1 namespace_controller.go:185] Namespace has been deleted subpath-8313\nI0718 20:02:59.114207       1 namespace_controller.go:185] Namespace has been deleted security-context-test-3173\nI0718 20:02:59.140598       1 pvc_protection_controller.go:291] \"PVC is unused\" PVC=\"ephemeral-5827/inline-volume-tester-knktf-my-volume-0\"\nI0718 20:02:59.149959       1 garbagecollector.go:471] \"Processing object\" object=\"ephemeral-5827/inline-volume-tester-knktf\" objectUID=9aebdd3c-8ab2-4b18-9157-091e2417cf4a kind=\"Pod\" virtual=false\nI0718 20:02:59.152421       1 garbagecollector.go:590] remove DeleteDependents finalizer for item [v1/Pod, namespace: ephemeral-5827, name: inline-volume-tester-knktf, uid: 9aebdd3c-8ab2-4b18-9157-091e2417cf4a]\nI0718 20:02:59.153348       1 pv_controller.go:640] volume \"pvc-dc473523-d38a-4923-8439-32b2189b8860\" is released and reclaim policy \"Delete\" will be executed\nI0718 20:02:59.159356       1 pv_controller.go:879] volume \"pvc-dc473523-d38a-4923-8439-32b2189b8860\" entered phase \"Released\"\nI0718 20:02:59.161474       1 pv_controller.go:1341] isVolumeReleased[pvc-dc473523-d38a-4923-8439-32b2189b8860]: volume is released\nI0718 20:02:59.188777       1 pv_controller_base.go:505] deletion of claim \"ephemeral-5827/inline-volume-tester-knktf-my-volume-0\" was already processed\nE0718 20:02:59.830365       1 reflector.go:138] k8s.io/client-go/metadata/metadatainformer/informer.go:90: Failed to watch *v1.PartialObjectMetadata: failed to list *v1.PartialObjectMetadata: the server could not find the requested resource\nE0718 20:02:59.966817       1 reflector.go:138] k8s.io/client-go/metadata/metadatainformer/informer.go:90: Failed to watch *v1.PartialObjectMetadata: failed to list *v1.PartialObjectMetadata: the server could not find the requested resource\nI0718 20:03:00.111558       1 pvc_protection_controller.go:291] \"PVC is unused\" PVC=\"volume-6367/pvc-5kgs5\"\nI0718 20:03:00.116257       1 event.go:291] \"Event occurred\" object=\"cronjob-9635/failed-jobs-history-limit\" kind=\"CronJob\" apiVersion=\"batch/v1\" type=\"Normal\" reason=\"SuccessfulCreate\" message=\"Created job failed-jobs-history-limit-27110643\"\nI0718 20:03:00.131775       1 cronjob_controllerv2.go:193] \"error cleaning up jobs\" cronjob=\"cronjob-9635/failed-jobs-history-limit\" resourceVersion=\"34993\" err=\"Operation cannot be fulfilled on cronjobs.batch \\\"failed-jobs-history-limit\\\": the object has been modified; please apply your changes to the latest version and try again\"\nE0718 20:03:00.131793       1 cronjob_controllerv2.go:154] error syncing CronJobController cronjob-9635/failed-jobs-history-limit, requeuing: Operation cannot be fulfilled on cronjobs.batch \"failed-jobs-history-limit\": the object has been modified; please apply your changes to the latest version and try again\nI0718 20:03:00.135749       1 event.go:291] \"Event occurred\" object=\"cronjob-9635/failed-jobs-history-limit-27110643\" kind=\"Job\" apiVersion=\"batch/v1\" type=\"Normal\" reason=\"SuccessfulCreate\" message=\"Created pod: failed-jobs-history-limit-27110643-fjbv4\"\nI0718 20:03:00.136699       1 pv_controller.go:640] volume \"local-m8msb\" is released and reclaim policy \"Retain\" will be executed\nI0718 20:03:00.151003       1 pv_controller.go:879] volume \"local-m8msb\" entered phase \"Released\"\nI0718 20:03:00.160729       1 pv_controller_base.go:505] deletion of claim \"volume-6367/pvc-5kgs5\" was already processed\nE0718 20:03:00.956880       1 tokens_controller.go:262] error synchronizing serviceaccount services-4562/default: secrets \"default-token-25mkc\" is forbidden: unable to create new content in namespace services-4562 because it is being terminated\nE0718 20:03:01.207116       1 namespace_controller.go:162] deletion of namespace services-4562 failed: unable to retrieve the complete list of server APIs: kubectl.example.com/v1: the server could not find the requested resource\nE0718 20:03:01.368876       1 namespace_controller.go:162] deletion of namespace services-4562 failed: unable to retrieve the complete list of server APIs: kubectl.example.com/v1: the server could not find the requested resource\nE0718 20:03:01.729186       1 namespace_controller.go:162] deletion of namespace services-4562 failed: unable to retrieve the complete list of server APIs: kubectl.example.com/v1: the server could not find the requested resource\nE0718 20:03:01.764724       1 namespace_controller.go:162] deletion of namespace ephemeral-599-8917 failed: unable to retrieve the complete list of server APIs: kubectl.example.com/v1: the server could not find the requested resource\nE0718 20:03:01.833134       1 namespace_controller.go:162] deletion of namespace kubelet-test-4596 failed: unable to retrieve the complete list of server APIs: kubectl.example.com/v1: the server could not find the requested resource\nE0718 20:03:01.931180       1 namespace_controller.go:162] deletion of namespace services-4562 failed: unable to retrieve the complete list of server APIs: kubectl.example.com/v1: the server could not find the requested resource\nE0718 20:03:01.938719       1 namespace_controller.go:162] deletion of namespace ephemeral-599-8917 failed: unable to retrieve the complete list of server APIs: kubectl.example.com/v1: the server could not find the requested resource\nE0718 20:03:02.002930       1 namespace_controller.go:162] deletion of namespace kubelet-test-4596 failed: unable to retrieve the complete list of server APIs: kubectl.example.com/v1: the server could not find the requested resource\nE0718 20:03:02.112314       1 reflector.go:138] k8s.io/client-go/metadata/metadatainformer/informer.go:90: Failed to watch *v1.PartialObjectMetadata: failed to list *v1.PartialObjectMetadata: the server could not find the requested resource\nI0718 20:03:02.367511       1 replica_set.go:559] \"Too few replicas\" replicaSet=\"webhook-9159/sample-webhook-deployment-78988fc6cd\" need=1 creating=1\nI0718 20:03:02.367656       1 event.go:291] \"Event occurred\" object=\"webhook-9159/sample-webhook-deployment\" kind=\"Deployment\" apiVersion=\"apps/v1\" type=\"Normal\" reason=\"ScalingReplicaSet\" message=\"Scaled up replica set sample-webhook-deployment-78988fc6cd to 1\"\nI0718 20:03:02.383959       1 deployment_controller.go:490] \"Error syncing deployment\" deployment=\"webhook-9159/sample-webhook-deployment\" err=\"Operation cannot be fulfilled on deployments.apps \\\"sample-webhook-deployment\\\": the object has been modified; please apply your changes to the latest version and try again\"\nI0718 20:03:02.387909       1 event.go:291] \"Event occurred\" object=\"webhook-9159/sample-webhook-deployment-78988fc6cd\" kind=\"ReplicaSet\" apiVersion=\"apps/v1\" type=\"Normal\" reason=\"SuccessfulCreate\" message=\"Created pod: sample-webhook-deployment-78988fc6cd-66fb6\"\nE0718 20:03:02.462037       1 namespace_controller.go:162] deletion of namespace services-4562 failed: unable to retrieve the complete list of server APIs: kubectl.example.com/v1: the server could not find the requested resource\nI0718 20:03:02.569146       1 reconciler.go:219] attacherDetacher.DetachVolume started for volume \"pvc-dc473523-d38a-4923-8439-32b2189b8860\" (UniqueName: \"kubernetes.io/csi/csi-hostpath-ephemeral-5827^ccc704c3-e802-11eb-8233-02fc2257e70c\") on node \"10.116.0.7\" \nI0718 20:03:02.572190       1 operation_generator.go:1483] Verified volume is safe to detach for volume \"pvc-dc473523-d38a-4923-8439-32b2189b8860\" (UniqueName: \"kubernetes.io/csi/csi-hostpath-ephemeral-5827^ccc704c3-e802-11eb-8233-02fc2257e70c\") on node \"10.116.0.7\" \nI0718 20:03:02.635105       1 replica_set.go:595] \"Too many replicas\" replicaSet=\"deployment-4244/webserver-789dfdff45\" need=0 deleting=1\nI0718 20:03:02.635151       1 replica_set.go:223] \"Found related ReplicaSets\" replicaSet=\"deployment-4244/webserver-789dfdff45\" relatedReplicaSets=[webserver-847dcfb7fb webserver-6584b976d5 webserver-789dfdff45 webserver-555c7ddb77]\nI0718 20:03:02.635234       1 controller_utils.go:602] \"Deleting pod\" controller=\"webserver-789dfdff45\" pod=\"deployment-4244/webserver-789dfdff45-ntnbk\"\nI0718 20:03:02.635733       1 event.go:291] \"Event occurred\" object=\"deployment-4244/webserver\" kind=\"Deployment\" apiVersion=\"apps/v1\" type=\"Normal\" reason=\"ScalingReplicaSet\" message=\"Scaled down replica set webserver-789dfdff45 to 0\"\nI0718 20:03:02.651590       1 replica_set.go:559] \"Too few replicas\" replicaSet=\"deployment-4244/webserver-555c7ddb77\" need=3 creating=1\nI0718 20:03:02.651731       1 event.go:291] \"Event occurred\" object=\"deployment-4244/webserver\" kind=\"Deployment\" apiVersion=\"apps/v1\" type=\"Normal\" reason=\"ScalingReplicaSet\" message=\"Scaled up replica set webserver-555c7ddb77 to 3\"\nI0718 20:03:02.664736       1 event.go:291] \"Event occurred\" object=\"deployment-4244/webserver-555c7ddb77\" kind=\"ReplicaSet\" apiVersion=\"apps/v1\" type=\"Normal\" reason=\"SuccessfulCreate\" message=\"Created pod: webserver-555c7ddb77-q9vqc\"\nI0718 20:03:02.666731       1 event.go:291] \"Event occurred\" object=\"deployment-4244/webserver-789dfdff45\" kind=\"ReplicaSet\" apiVersion=\"apps/v1\" type=\"Normal\" reason=\"SuccessfulDelete\" message=\"Deleted pod: webserver-789dfdff45-ntnbk\"\nE0718 20:03:02.683223       1 namespace_controller.go:162] deletion of namespace ephemeral-599-8917 failed: unable to retrieve the complete list of server APIs: kubectl.example.com/v1: the server could not find the requested resource\nE0718 20:03:02.739373       1 namespace_controller.go:162] deletion of namespace port-forwarding-3146 failed: unable to retrieve the complete list of server APIs: kubectl.example.com/v1: the server could not find the requested resource\nE0718 20:03:02.756146       1 namespace_controller.go:162] deletion of namespace kubelet-test-4596 failed: unable to retrieve the complete list of server APIs: kubectl.example.com/v1: the server could not find the requested resource\nE0718 20:03:02.844612       1 namespace_controller.go:162] deletion of namespace services-4562 failed: unable to retrieve the complete list of server APIs: kubectl.example.com/v1: the server could not find the requested resource\nE0718 20:03:02.891153       1 namespace_controller.go:162] deletion of namespace ephemeral-599-8917 failed: unable to retrieve the complete list of server APIs: kubectl.example.com/v1: the server could not find the requested resource\nE0718 20:03:02.934089       1 namespace_controller.go:162] deletion of namespace port-forwarding-3146 failed: unable to retrieve the complete list of server APIs: kubectl.example.com/v1: the server could not find the requested resource\nE0718 20:03:02.985531       1 namespace_controller.go:162] deletion of namespace kubelet-test-4596 failed: unable to retrieve the complete list of server APIs: kubectl.example.com/v1: the server could not find the requested resource\nE0718 20:03:03.121045       1 namespace_controller.go:162] deletion of namespace ephemeral-599-8917 failed: unable to retrieve the complete list of server APIs: kubectl.example.com/v1: the server could not find the requested resource\nI0718 20:03:03.125855       1 operation_generator.go:483] DetachVolume.Detach succeeded for volume \"pvc-dc473523-d38a-4923-8439-32b2189b8860\" (UniqueName: \"kubernetes.io/csi/csi-hostpath-ephemeral-5827^ccc704c3-e802-11eb-8233-02fc2257e70c\") on node \"10.116.0.7\" \nE0718 20:03:03.137881       1 namespace_controller.go:162] deletion of namespace port-forwarding-3146 failed: unable to retrieve the complete list of server APIs: kubectl.example.com/v1: the server could not find the requested resource\nE0718 20:03:03.195443       1 namespace_controller.go:162] deletion of namespace services-4562 failed: unable to retrieve the complete list of server APIs: kubectl.example.com/v1: the server could not find the requested resource\nE0718 20:03:03.260481       1 namespace_controller.go:162] deletion of namespace kubelet-test-4596 failed: unable to retrieve the complete list of server APIs: kubectl.example.com/v1: the server could not find the requested resource\nE0718 20:03:03.328601       1 namespace_controller.go:162] deletion of namespace port-forwarding-3146 failed: unable to retrieve the complete list of server APIs: kubectl.example.com/v1: the server could not find the requested resource\nE0718 20:03:03.364331       1 namespace_controller.go:162] deletion of namespace ephemeral-599-8917 failed: unable to retrieve the complete list of server APIs: kubectl.example.com/v1: the server could not find the requested resource\nI0718 20:03:03.419129       1 pvc_protection_controller.go:291] \"PVC is unused\" PVC=\"provisioning-3398/pvc-r869k\"\nI0718 20:03:03.429942       1 pv_controller.go:640] volume \"local-jmrl9\" is released and reclaim policy \"Retain\" will be executed\nI0718 20:03:03.435367       1 pv_controller.go:879] volume \"local-jmrl9\" entered phase \"Released\"\nI0718 20:03:03.457252       1 pv_controller_base.go:505] deletion of claim \"provisioning-3398/pvc-r869k\" was already processed\nE0718 20:03:03.503842       1 reflector.go:138] k8s.io/client-go/metadata/metadatainformer/informer.go:90: Failed to watch *v1.PartialObjectMetadata: failed to list *v1.PartialObjectMetadata: the server could not find the requested resource\nI0718 20:03:03.507317       1 pv_controller.go:879] volume \"local-pv2gzbg\" entered phase \"Available\"\nI0718 20:03:03.536399       1 pv_controller.go:930] claim \"persistent-local-volumes-test-7580/pvc-59wj4\" bound to volume \"local-pv2gzbg\"\nI0718 20:03:03.549073       1 pv_controller.go:879] volume \"local-pv2gzbg\" entered phase \"Bound\"\nI0718 20:03:03.549095       1 pv_controller.go:982] volume \"local-pv2gzbg\" bound to claim \"persistent-local-volumes-test-7580/pvc-59wj4\"\nI0718 20:03:03.571973       1 pv_controller.go:823] claim \"persistent-local-volumes-test-7580/pvc-59wj4\" entered phase \"Bound\"\nE0718 20:03:03.589525       1 namespace_controller.go:162] deletion of namespace port-forwarding-3146 failed: unable to retrieve the complete list of server APIs: kubectl.example.com/v1: the server could not find the requested resource\nI0718 20:03:03.622680       1 replica_set.go:559] \"Too few replicas\" replicaSet=\"deployment-4244/webserver-555c7ddb77\" need=5 creating=2\nI0718 20:03:03.622834       1 event.go:291] \"Event occurred\" object=\"deployment-4244/webserver\" kind=\"Deployment\" apiVersion=\"apps/v1\" type=\"Normal\" reason=\"ScalingReplicaSet\" message=\"Scaled up replica set webserver-555c7ddb77 to 5\"\nI0718 20:03:03.625595       1 namespace_controller.go:185] Namespace has been deleted provisioning-7043-7873\nI0718 20:03:03.631671       1 event.go:291] \"Event occurred\" object=\"deployment-4244/webserver-555c7ddb77\" kind=\"ReplicaSet\" apiVersion=\"apps/v1\" type=\"Normal\" reason=\"SuccessfulCreate\" message=\"Created pod: webserver-555c7ddb77-pmnmd\"\nI0718 20:03:03.641953       1 event.go:291] \"Event occurred\" object=\"deployment-4244/webserver-555c7ddb77\" kind=\"ReplicaSet\" apiVersion=\"apps/v1\" type=\"Normal\" reason=\"SuccessfulCreate\" message=\"Created pod: webserver-555c7ddb77-xxbq7\"\nE0718 20:03:03.666431       1 namespace_controller.go:162] deletion of namespace kubelet-test-4596 failed: unable to retrieve the complete list of server APIs: kubectl.example.com/v1: the server could not find the requested resource\nI0718 20:03:03.674313       1 deployment_controller.go:490] \"Error syncing deployment\" deployment=\"deployment-4244/webserver\" err=\"Operation cannot be fulfilled on deployments.apps \\\"webserver\\\": the object has been modified; please apply your changes to the latest version and try again\"\nI0718 20:03:03.708979       1 replica_set.go:559] \"Too few replicas\" replicaSet=\"deployment-4244/webserver-555c7ddb77\" need=5 creating=1\nI0718 20:03:03.716283       1 event.go:291] \"Event occurred\" object=\"deployment-4244/webserver-555c7ddb77\" kind=\"ReplicaSet\" apiVersion=\"apps/v1\" type=\"Normal\" reason=\"SuccessfulCreate\" message=\"Created pod: webserver-555c7ddb77-klb9c\"\nI0718 20:03:03.756963       1 replica_set.go:559] \"Too few replicas\" replicaSet=\"deployment-4244/webserver-555c7ddb77\" need=5 creating=1\nI0718 20:03:03.767572       1 event.go:291] \"Event occurred\" object=\"deployment-4244/webserver-555c7ddb77\" kind=\"ReplicaSet\" apiVersion=\"apps/v1\" type=\"Normal\" reason=\"SuccessfulCreate\" message=\"Created pod: webserver-555c7ddb77-27b47\"\nI0718 20:03:03.807239       1 replica_set.go:559] \"Too few replicas\" replicaSet=\"deployment-4244/webserver-555c7ddb77\" need=5 creating=1\nI0718 20:03:03.813814       1 event.go:291] \"Event occurred\" object=\"deployment-4244/webserver-555c7ddb77\" kind=\"ReplicaSet\" apiVersion=\"apps/v1\" type=\"Normal\" reason=\"SuccessfulCreate\" message=\"Created pod: webserver-555c7ddb77-qq7kz\"\nE0718 20:03:03.853089       1 namespace_controller.go:162] deletion of namespace ephemeral-599-8917 failed: unable to retrieve the complete list of server APIs: kubectl.example.com/v1: the server could not find the requested resource\nE0718 20:03:03.853992       1 namespace_controller.go:162] deletion of namespace services-4562 failed: unable to retrieve the complete list of server APIs: kubectl.example.com/v1: the server could not find the requested resource\nI0718 20:03:03.883099       1 replica_set.go:559] \"Too few replicas\" replicaSet=\"deployment-4244/webserver-555c7ddb77\" need=6 creating=1\nI0718 20:03:03.883722       1 event.go:291] \"Event occurred\" object=\"deployment-4244/webserver\" kind=\"Deployment\" apiVersion=\"apps/v1\" type=\"Normal\" reason=\"ScalingReplicaSet\" message=\"Scaled up replica set webserver-555c7ddb77 to 6\"\nI0718 20:03:03.894983       1 event.go:291] \"Event occurred\" object=\"deployment-4244/webserver-555c7ddb77\" kind=\"ReplicaSet\" apiVersion=\"apps/v1\" type=\"Normal\" reason=\"SuccessfulCreate\" message=\"Created pod: webserver-555c7ddb77-jqtbl\"\nI0718 20:03:03.930132       1 replica_set.go:559] \"Too few replicas\" replicaSet=\"apply-3616/deployment-shared-map-item-removal-55649fd747\" need=3 creating=3\nI0718 20:03:03.931068       1 event.go:291] \"Event occurred\" object=\"apply-3616/deployment-shared-map-item-removal\" kind=\"Deployment\" apiVersion=\"apps/v1\" type=\"Normal\" reason=\"ScalingReplicaSet\" message=\"Scaled up replica set deployment-shared-map-item-removal-55649fd747 to 3\"\nI0718 20:03:03.942281       1 event.go:291] \"Event occurred\" object=\"apply-3616/deployment-shared-map-item-removal-55649fd747\" kind=\"ReplicaSet\" apiVersion=\"apps/v1\" type=\"Normal\" reason=\"SuccessfulCreate\" message=\"Created pod: deployment-shared-map-item-removal-55649fd747-qr775\"\nI0718 20:03:03.951322       1 event.go:291] \"Event occurred\" object=\"apply-3616/deployment-shared-map-item-removal-55649fd747\" kind=\"ReplicaSet\" apiVersion=\"apps/v1\" type=\"Normal\" reason=\"SuccessfulCreate\" message=\"Created pod: deployment-shared-map-item-removal-55649fd747-l58c7\"\nI0718 20:03:03.953069       1 event.go:291] \"Event occurred\" object=\"apply-3616/deployment-shared-map-item-removal-55649fd747\" kind=\"ReplicaSet\" apiVersion=\"apps/v1\" type=\"Normal\" reason=\"SuccessfulCreate\" message=\"Created pod: deployment-shared-map-item-removal-55649fd747-9zpp6\"\nI0718 20:03:03.953205       1 deployment_controller.go:490] \"Error syncing deployment\" deployment=\"apply-3616/deployment-shared-map-item-removal\" err=\"Operation cannot be fulfilled on deployments.apps \\\"deployment-shared-map-item-removal\\\": the object has been modified; please apply your changes to the latest version and try again\"\nE0718 20:03:03.990410       1 namespace_controller.go:162] deletion of namespace port-forwarding-3146 failed: unable to retrieve the complete list of server APIs: kubectl.example.com/v1: the server could not find the requested resource\nI0718 20:03:04.037982       1 event.go:291] \"Event occurred\" object=\"csi-mock-volumes-1746-693/csi-mockplugin\" kind=\"StatefulSet\" apiVersion=\"apps/v1\" type=\"Normal\" reason=\"SuccessfulCreate\" message=\"create Pod csi-mockplugin-0 in StatefulSet csi-mockplugin successful\"\nI0718 20:03:04.101579       1 event.go:291] \"Event occurred\" object=\"csi-mock-volumes-1746-693/csi-mockplugin-attacher\" kind=\"StatefulSet\" apiVersion=\"apps/v1\" type=\"Normal\" reason=\"SuccessfulCreate\" message=\"create Pod csi-mockplugin-attacher-0 in StatefulSet csi-mockplugin-attacher successful\"\nI0718 20:03:04.191799       1 expand_controller.go:289] Ignoring the PVC \"csi-mock-volumes-9933/pvc-hm6bq\" (uid: \"dd7fe695-934e-4f09-8209-37cfbf1678ed\") : didn't find a plugin capable of expanding the volume; waiting for an external controller to process this PVC.\nI0718 20:03:04.191900       1 event.go:291] \"Event occurred\" object=\"csi-mock-volumes-9933/pvc-hm6bq\" kind=\"PersistentVolumeClaim\" apiVersion=\"v1\" type=\"Warning\" reason=\"ExternalExpanding\" message=\"Ignoring the PVC: didn't find a plugin capable of expanding the volume; waiting for an external controller to process this PVC.\"\nE0718 20:03:04.201474       1 namespace_controller.go:162] deletion of namespace kubelet-test-4596 failed: [unable to retrieve the complete list of server APIs: kubectl.example.com/v1: the server could not find the requested resource, unexpected items still remain in namespace: kubelet-test-4596 for gvr: /v1, Resource=pods]\nE0718 20:03:04.330558       1 namespace_controller.go:162] deletion of namespace port-forwarding-3146 failed: unable to retrieve the complete list of server APIs: kubectl.example.com/v1: the server could not find the requested resource\nE0718 20:03:04.354630       1 namespace_controller.go:162] deletion of namespace ephemeral-599-8917 failed: unable to retrieve the complete list of server APIs: kubectl.example.com/v1: the server could not find the requested resource\nE0718 20:03:04.563689       1 namespace_controller.go:162] deletion of namespace deployment-6861 failed: unable to retrieve the complete list of server APIs: kubectl.example.com/v1: the server could not find the requested resource\nE0718 20:03:04.662919       1 namespace_controller.go:162] deletion of namespace services-4562 failed: unable to retrieve the complete list of server APIs: kubectl.example.com/v1: the server could not find the requested resource\nE0718 20:03:04.724573       1 namespace_controller.go:162] deletion of namespace deployment-6861 failed: unable to retrieve the complete list of server APIs: kubectl.example.com/v1: the server could not find the requested resource\nE0718 20:03:04.830362       1 namespace_controller.go:162] deletion of namespace port-forwarding-3146 failed: unable to retrieve the complete list of server APIs: kubectl.example.com/v1: the server could not find the requested resource\nE0718 20:03:04.952142       1 tokens_controller.go:262] error synchronizing serviceaccount ephemeral-5827/default: secrets \"default-token-hpxdd\" is forbidden: unable to create new content in namespace ephemeral-5827 because it is being terminated\nE0718 20:03:05.053049       1 namespace_controller.go:162] deletion of namespace kubelet-test-4596 failed: unexpected items still remain in namespace: kubelet-test-4596 for gvr: /v1, Resource=pods\nI0718 20:03:06.025466       1 replica_set.go:559] \"Too few replicas\" replicaSet=\"deployment-4244/webserver-555c7ddb77\" need=7 creating=1\nI0718 20:03:06.025583       1 event.go:291] \"Event occurred\" object=\"deployment-4244/webserver\" kind=\"Deployment\" apiVersion=\"apps/v1\" type=\"Normal\" reason=\"ScalingReplicaSet\" message=\"Scaled up replica set webserver-555c7ddb77 to 7\"\nI0718 20:03:06.039429       1 event.go:291] \"Event occurred\" object=\"deployment-4244/webserver-555c7ddb77\" kind=\"ReplicaSet\" apiVersion=\"apps/v1\" type=\"Normal\" reason=\"SuccessfulCreate\" message=\"Created pod: webserver-555c7ddb77-hswkj\"\nI0718 20:03:06.065279       1 replica_set.go:559] \"Too few replicas\" replicaSet=\"apply-3616/deployment-shared-map-item-removal-55649fd747\" need=4 creating=1\nI0718 20:03:06.065485       1 event.go:291] \"Event occurred\" object=\"apply-3616/deployment-shared-map-item-removal\" kind=\"Deployment\" apiVersion=\"apps/v1\" type=\"Normal\" reason=\"ScalingReplicaSet\" message=\"Scaled up replica set deployment-shared-map-item-removal-55649fd747 to 4\"\nI0718 20:03:06.073072       1 event.go:291] \"Event occurred\" object=\"apply-3616/deployment-shared-map-item-removal-55649fd747\" kind=\"ReplicaSet\" apiVersion=\"apps/v1\" type=\"Normal\" reason=\"SuccessfulCreate\" message=\"Created pod: deployment-shared-map-item-removal-55649fd747-vzr89\"\nI0718 20:03:06.130933       1 controller_utils.go:602] \"Deleting pod\" controller=\"failed-jobs-history-limit-27110643\" pod=\"cronjob-9635/failed-jobs-history-limit-27110643-fjbv4\"\nI0718 20:03:06.139691       1 event.go:291] \"Event occurred\" object=\"cronjob-9635/failed-jobs-history-limit-27110643\" kind=\"Job\" apiVersion=\"batch/v1\" type=\"Normal\" reason=\"SuccessfulDelete\" message=\"Deleted pod: failed-jobs-history-limit-27110643-fjbv4\"\nI0718 20:03:06.139721       1 event.go:291] \"Event occurred\" object=\"cronjob-9635/failed-jobs-history-limit-27110643\" kind=\"Job\" apiVersion=\"batch/v1\" type=\"Warning\" reason=\"BackoffLimitExceeded\" message=\"Job has reached the specified backoff limit\"\nI0718 20:03:06.147292       1 event.go:291] \"Event occurred\" object=\"cronjob-9635/failed-jobs-history-limit\" kind=\"CronJob\" apiVersion=\"batch/v1\" type=\"Normal\" reason=\"SawCompletedJob\" message=\"Saw completed job: failed-jobs-history-limit-27110643, status: Failed\"\nI0718 20:03:06.275470       1 garbagecollector.go:471] \"Processing object\" object=\"apply-3616/deployment-shared-map-item-removal-55649fd747\" objectUID=3fbdf6d3-961a-4cca-8cfb-8d90d8b1e917 kind=\"ReplicaSet\" virtual=false\nI0718 20:03:06.275722       1 deployment_controller.go:583] \"Deployment has been deleted\" deployment=\"apply-3616/deployment-shared-map-item-removal\"\nI0718 20:03:06.278736       1 garbagecollector.go:580] \"Deleting object\" object=\"apply-3616/deployment-shared-map-item-removal-55649fd747\" objectUID=3fbdf6d3-961a-4cca-8cfb-8d90d8b1e917 kind=\"ReplicaSet\" propagationPolicy=Background\nI0718 20:03:06.284056       1 garbagecollector.go:471] \"Processing object\" object=\"apply-3616/deployment-shared-map-item-removal-55649fd747-qr775\" objectUID=d7a4253b-84eb-4af3-b2fd-a50df4b6fa88 kind=\"Pod\" virtual=false\nI0718 20:03:06.284103       1 garbagecollector.go:471] \"Processing object\" object=\"apply-3616/deployment-shared-map-item-removal-55649fd747-l58c7\" objectUID=649d97f2-bf2a-49a8-a69b-00ac1f05026f kind=\"Pod\" virtual=false\nI0718 20:03:06.284158       1 garbagecollector.go:471] \"Processing object\" object=\"apply-3616/deployment-shared-map-item-removal-55649fd747-9zpp6\" objectUID=fae53663-3137-4544-a241-7dee7193bb48 kind=\"Pod\" virtual=false\nI0718 20:03:06.284246       1 garbagecollector.go:471] \"Processing object\" object=\"apply-3616/deployment-shared-map-item-removal-55649fd747-vzr89\" objectUID=bb51c7dd-3ac3-4f7b-aec2-b1e0acb43332 kind=\"Pod\" virtual=false\nI0718 20:03:06.286219       1 garbagecollector.go:580] \"Deleting object\" object=\"apply-3616/deployment-shared-map-item-removal-55649fd747-qr775\" objectUID=d7a4253b-84eb-4af3-b2fd-a50df4b6fa88 kind=\"Pod\" propagationPolicy=Background\nI0718 20:03:06.287070       1 garbagecollector.go:580] \"Deleting object\" object=\"apply-3616/deployment-shared-map-item-removal-55649fd747-9zpp6\" objectUID=fae53663-3137-4544-a241-7dee7193bb48 kind=\"Pod\" propagationPolicy=Background\nI0718 20:03:06.287165       1 garbagecollector.go:580] \"Deleting object\" object=\"apply-3616/deployment-shared-map-item-removal-55649fd747-l58c7\" objectUID=649d97f2-bf2a-49a8-a69b-00ac1f05026f kind=\"Pod\" propagationPolicy=Background\nI0718 20:03:06.287166       1 garbagecollector.go:580] \"Deleting object\" object=\"apply-3616/deployment-shared-map-item-removal-55649fd747-vzr89\" objectUID=bb51c7dd-3ac3-4f7b-aec2-b1e0acb43332 kind=\"Pod\" propagationPolicy=Background\nE0718 20:03:06.336683       1 reflector.go:138] k8s.io/client-go/metadata/metadatainformer/informer.go:90: