This job view page is being replaced by Spyglass soon. Check out the new job view.
ResultFAILURE
Tests 0 failed / 1 succeeded
Started2022-07-29 14:01
Elapsed15m35s
Revisionmaster

No Test Failures!


Show 1 Passed Tests

Error lines from build-log.txt

... skipping 212 lines ...
I0729 14:02:30.784464    6210 app.go:63] pass rundir-in-artifacts flag True for RunDir to be part of Artifacts
I0729 14:02:30.784488    6210 app.go:65] RunDir for this run: "/home/prow/go/src/k8s.io/kops/_rundir/d1d52443-0f46-11ed-b6e4-d28c295d3c4e"
I0729 14:02:30.836003    6210 app.go:129] ID for this run: "d1d52443-0f46-11ed-b6e4-d28c295d3c4e"
I0729 14:02:30.836352    6210 local.go:42] ⚙️ ssh-keygen -t ed25519 -N  -q -f /tmp/kops/e2e-414ec24bfc-41f44.test-cncf-aws.k8s.io/id_ed25519
I0729 14:02:30.847020    6210 dumplogs.go:45] /tmp/kops.Gc2Stcgpy toolbox dump --name e2e-414ec24bfc-41f44.test-cncf-aws.k8s.io --dir /logs/artifacts --private-key /tmp/kops/e2e-414ec24bfc-41f44.test-cncf-aws.k8s.io/id_ed25519 --ssh-user ubuntu
I0729 14:02:30.847498    6210 local.go:42] ⚙️ /tmp/kops.Gc2Stcgpy toolbox dump --name e2e-414ec24bfc-41f44.test-cncf-aws.k8s.io --dir /logs/artifacts --private-key /tmp/kops/e2e-414ec24bfc-41f44.test-cncf-aws.k8s.io/id_ed25519 --ssh-user ubuntu
W0729 14:02:31.397443    6210 down.go:34] Dumping cluster logs at the start of Down() failed: exit status 1
I0729 14:02:31.397497    6210 down.go:48] /tmp/kops.Gc2Stcgpy delete cluster --name e2e-414ec24bfc-41f44.test-cncf-aws.k8s.io --yes
I0729 14:02:31.397508    6210 local.go:42] ⚙️ /tmp/kops.Gc2Stcgpy delete cluster --name e2e-414ec24bfc-41f44.test-cncf-aws.k8s.io --yes
I0729 14:02:31.419771    6235 featureflag.go:162] FeatureFlag "SpecOverrideFlag"=true
I0729 14:02:31.419888    6235 featureflag.go:162] FeatureFlag "SpecOverrideFlag"=true
Error: error reading cluster configuration: Cluster.kops.k8s.io "e2e-414ec24bfc-41f44.test-cncf-aws.k8s.io" not found
Error: exit status 1
+ echo 'kubetest2 down failed'
kubetest2 down failed
+ [[ v == \v ]]
+ KOPS_BASE_URL=
++ kops-download-release v1.24.0
++ local kops
+++ mktemp -t kops.XXXXXXXXX
++ kops=/tmp/kops.bMEwXgKA5
... skipping 10 lines ...
I0729 14:02:34.897757    6269 app.go:63] pass rundir-in-artifacts flag True for RunDir to be part of Artifacts
I0729 14:02:34.897779    6269 app.go:65] RunDir for this run: "/home/prow/go/src/k8s.io/kops/_rundir/d1d52443-0f46-11ed-b6e4-d28c295d3c4e"
I0729 14:02:34.901093    6269 app.go:129] ID for this run: "d1d52443-0f46-11ed-b6e4-d28c295d3c4e"
I0729 14:02:34.901156    6269 up.go:44] Cleaning up any leaked resources from previous cluster
I0729 14:02:34.901185    6269 dumplogs.go:45] /tmp/kops.bMEwXgKA5 toolbox dump --name e2e-414ec24bfc-41f44.test-cncf-aws.k8s.io --dir /logs/artifacts --private-key /tmp/kops/e2e-414ec24bfc-41f44.test-cncf-aws.k8s.io/id_ed25519 --ssh-user ubuntu
I0729 14:02:34.901222    6269 local.go:42] ⚙️ /tmp/kops.bMEwXgKA5 toolbox dump --name e2e-414ec24bfc-41f44.test-cncf-aws.k8s.io --dir /logs/artifacts --private-key /tmp/kops/e2e-414ec24bfc-41f44.test-cncf-aws.k8s.io/id_ed25519 --ssh-user ubuntu
W0729 14:02:35.412557    6269 down.go:34] Dumping cluster logs at the start of Down() failed: exit status 1
I0729 14:02:35.412597    6269 down.go:48] /tmp/kops.bMEwXgKA5 delete cluster --name e2e-414ec24bfc-41f44.test-cncf-aws.k8s.io --yes
I0729 14:02:35.412608    6269 local.go:42] ⚙️ /tmp/kops.bMEwXgKA5 delete cluster --name e2e-414ec24bfc-41f44.test-cncf-aws.k8s.io --yes
I0729 14:02:35.433821    6290 featureflag.go:162] FeatureFlag "SpecOverrideFlag"=true
I0729 14:02:35.433936    6290 featureflag.go:162] FeatureFlag "SpecOverrideFlag"=true
Error: error reading cluster configuration: Cluster.kops.k8s.io "e2e-414ec24bfc-41f44.test-cncf-aws.k8s.io" not found
I0729 14:02:35.896635    6269 http.go:37] curl http://metadata.google.internal/computeMetadata/v1/instance/network-interfaces/0/access-configs/0/external-ip
2022/07/29 14:02:35 failed to get external ip from metadata service: http://metadata.google.internal/computeMetadata/v1/instance/network-interfaces/0/access-configs/0/external-ip returned 404
I0729 14:02:35.908032    6269 http.go:37] curl https://ip.jsb.workers.dev
I0729 14:02:36.023996    6269 template.go:58] /tmp/kops.bMEwXgKA5 toolbox template --template tests/e2e/templates/many-addons.yaml.tmpl --output /tmp/kops-template2369809461/manifest.yaml --values /tmp/kops-template2369809461/values.yaml --name e2e-414ec24bfc-41f44.test-cncf-aws.k8s.io
I0729 14:02:36.024036    6269 local.go:42] ⚙️ /tmp/kops.bMEwXgKA5 toolbox template --template tests/e2e/templates/many-addons.yaml.tmpl --output /tmp/kops-template2369809461/manifest.yaml --values /tmp/kops-template2369809461/values.yaml --name e2e-414ec24bfc-41f44.test-cncf-aws.k8s.io
I0729 14:02:36.044233    6302 featureflag.go:162] FeatureFlag "SpecOverrideFlag"=true
I0729 14:02:36.044319    6302 featureflag.go:162] FeatureFlag "SpecOverrideFlag"=true
I0729 14:02:36.146745    6269 create.go:33] /tmp/kops.bMEwXgKA5 create --filename /tmp/kops-template2369809461/manifest.yaml --name e2e-414ec24bfc-41f44.test-cncf-aws.k8s.io
... skipping 66 lines ...

NODE STATUS
NAME	ROLE	READY

VALIDATION ERRORS
KIND	NAME		MESSAGE
dns	apiserver	Validation Failed

The dns-controller Kubernetes deployment has not updated the Kubernetes cluster's API DNS entry to the correct IP address.  The API DNS IP address is the placeholder address that kops creates: 203.0.113.123.  Please wait about 5-10 minutes for a master to start, dns-controller to launch, and DNS to propagate.  The protokube container and dns-controller deployment logs may contain more diagnostic information.  Etcd and the API DNS entries must be updated for a kops Kubernetes cluster to start.

Validation Failed
W0729 14:03:20.640400    6342 validate_cluster.go:232] (will retry): cluster not yet healthy
INSTANCE GROUPS
NAME			ROLE	MACHINETYPE	MIN	MAX	SUBNETS
master-eu-west-3a	Master	c5.large	1	1	eu-west-3a
nodes-eu-west-3a	Node	t3.medium	4	4	eu-west-3a

NODE STATUS
NAME	ROLE	READY

VALIDATION ERRORS
KIND	NAME		MESSAGE
dns	apiserver	Validation Failed

The dns-controller Kubernetes deployment has not updated the Kubernetes cluster's API DNS entry to the correct IP address.  The API DNS IP address is the placeholder address that kops creates: 203.0.113.123.  Please wait about 5-10 minutes for a master to start, dns-controller to launch, and DNS to propagate.  The protokube container and dns-controller deployment logs may contain more diagnostic information.  Etcd and the API DNS entries must be updated for a kops Kubernetes cluster to start.

Validation Failed
W0729 14:03:30.683040    6342 validate_cluster.go:232] (will retry): cluster not yet healthy
INSTANCE GROUPS
NAME			ROLE	MACHINETYPE	MIN	MAX	SUBNETS
master-eu-west-3a	Master	c5.large	1	1	eu-west-3a
nodes-eu-west-3a	Node	t3.medium	4	4	eu-west-3a

NODE STATUS
NAME	ROLE	READY

VALIDATION ERRORS
KIND	NAME		MESSAGE
dns	apiserver	Validation Failed

The dns-controller Kubernetes deployment has not updated the Kubernetes cluster's API DNS entry to the correct IP address.  The API DNS IP address is the placeholder address that kops creates: 203.0.113.123.  Please wait about 5-10 minutes for a master to start, dns-controller to launch, and DNS to propagate.  The protokube container and dns-controller deployment logs may contain more diagnostic information.  Etcd and the API DNS entries must be updated for a kops Kubernetes cluster to start.

Validation Failed
W0729 14:03:40.722744    6342 validate_cluster.go:232] (will retry): cluster not yet healthy
INSTANCE GROUPS
NAME			ROLE	MACHINETYPE	MIN	MAX	SUBNETS
master-eu-west-3a	Master	c5.large	1	1	eu-west-3a
nodes-eu-west-3a	Node	t3.medium	4	4	eu-west-3a

NODE STATUS
NAME	ROLE	READY

VALIDATION ERRORS
KIND	NAME		MESSAGE
dns	apiserver	Validation Failed

The dns-controller Kubernetes deployment has not updated the Kubernetes cluster's API DNS entry to the correct IP address.  The API DNS IP address is the placeholder address that kops creates: 203.0.113.123.  Please wait about 5-10 minutes for a master to start, dns-controller to launch, and DNS to propagate.  The protokube container and dns-controller deployment logs may contain more diagnostic information.  Etcd and the API DNS entries must be updated for a kops Kubernetes cluster to start.

Validation Failed
W0729 14:03:50.763834    6342 validate_cluster.go:232] (will retry): cluster not yet healthy
INSTANCE GROUPS
NAME			ROLE	MACHINETYPE	MIN	MAX	SUBNETS
master-eu-west-3a	Master	c5.large	1	1	eu-west-3a
nodes-eu-west-3a	Node	t3.medium	4	4	eu-west-3a

NODE STATUS
NAME	ROLE	READY

VALIDATION ERRORS
KIND	NAME		MESSAGE
dns	apiserver	Validation Failed

The dns-controller Kubernetes deployment has not updated the Kubernetes cluster's API DNS entry to the correct IP address.  The API DNS IP address is the placeholder address that kops creates: 203.0.113.123.  Please wait about 5-10 minutes for a master to start, dns-controller to launch, and DNS to propagate.  The protokube container and dns-controller deployment logs may contain more diagnostic information.  Etcd and the API DNS entries must be updated for a kops Kubernetes cluster to start.

Validation Failed
W0729 14:04:00.800156    6342 validate_cluster.go:232] (will retry): cluster not yet healthy
INSTANCE GROUPS
NAME			ROLE	MACHINETYPE	MIN	MAX	SUBNETS
master-eu-west-3a	Master	c5.large	1	1	eu-west-3a
nodes-eu-west-3a	Node	t3.medium	4	4	eu-west-3a

NODE STATUS
NAME	ROLE	READY

VALIDATION ERRORS
KIND	NAME		MESSAGE
dns	apiserver	Validation Failed

The dns-controller Kubernetes deployment has not updated the Kubernetes cluster's API DNS entry to the correct IP address.  The API DNS IP address is the placeholder address that kops creates: 203.0.113.123.  Please wait about 5-10 minutes for a master to start, dns-controller to launch, and DNS to propagate.  The protokube container and dns-controller deployment logs may contain more diagnostic information.  Etcd and the API DNS entries must be updated for a kops Kubernetes cluster to start.

Validation Failed
W0729 14:04:10.837735    6342 validate_cluster.go:232] (will retry): cluster not yet healthy
INSTANCE GROUPS
NAME			ROLE	MACHINETYPE	MIN	MAX	SUBNETS
master-eu-west-3a	Master	c5.large	1	1	eu-west-3a
nodes-eu-west-3a	Node	t3.medium	4	4	eu-west-3a

NODE STATUS
NAME	ROLE	READY

VALIDATION ERRORS
KIND	NAME		MESSAGE
dns	apiserver	Validation Failed

The dns-controller Kubernetes deployment has not updated the Kubernetes cluster's API DNS entry to the correct IP address.  The API DNS IP address is the placeholder address that kops creates: 203.0.113.123.  Please wait about 5-10 minutes for a master to start, dns-controller to launch, and DNS to propagate.  The protokube container and dns-controller deployment logs may contain more diagnostic information.  Etcd and the API DNS entries must be updated for a kops Kubernetes cluster to start.

Validation Failed
W0729 14:04:20.874399    6342 validate_cluster.go:232] (will retry): cluster not yet healthy
INSTANCE GROUPS
NAME			ROLE	MACHINETYPE	MIN	MAX	SUBNETS
master-eu-west-3a	Master	c5.large	1	1	eu-west-3a
nodes-eu-west-3a	Node	t3.medium	4	4	eu-west-3a

NODE STATUS
NAME	ROLE	READY

VALIDATION ERRORS
KIND	NAME		MESSAGE
dns	apiserver	Validation Failed

The dns-controller Kubernetes deployment has not updated the Kubernetes cluster's API DNS entry to the correct IP address.  The API DNS IP address is the placeholder address that kops creates: 203.0.113.123.  Please wait about 5-10 minutes for a master to start, dns-controller to launch, and DNS to propagate.  The protokube container and dns-controller deployment logs may contain more diagnostic information.  Etcd and the API DNS entries must be updated for a kops Kubernetes cluster to start.

Validation Failed
W0729 14:04:30.917267    6342 validate_cluster.go:232] (will retry): cluster not yet healthy
INSTANCE GROUPS
NAME			ROLE	MACHINETYPE	MIN	MAX	SUBNETS
master-eu-west-3a	Master	c5.large	1	1	eu-west-3a
nodes-eu-west-3a	Node	t3.medium	4	4	eu-west-3a

NODE STATUS
NAME	ROLE	READY

VALIDATION ERRORS
KIND	NAME		MESSAGE
dns	apiserver	Validation Failed

The dns-controller Kubernetes deployment has not updated the Kubernetes cluster's API DNS entry to the correct IP address.  The API DNS IP address is the placeholder address that kops creates: 203.0.113.123.  Please wait about 5-10 minutes for a master to start, dns-controller to launch, and DNS to propagate.  The protokube container and dns-controller deployment logs may contain more diagnostic information.  Etcd and the API DNS entries must be updated for a kops Kubernetes cluster to start.

Validation Failed
W0729 14:04:40.969725    6342 validate_cluster.go:232] (will retry): cluster not yet healthy
INSTANCE GROUPS
NAME			ROLE	MACHINETYPE	MIN	MAX	SUBNETS
master-eu-west-3a	Master	c5.large	1	1	eu-west-3a
nodes-eu-west-3a	Node	t3.medium	4	4	eu-west-3a

NODE STATUS
NAME	ROLE	READY

VALIDATION ERRORS
KIND	NAME		MESSAGE
dns	apiserver	Validation Failed

The dns-controller Kubernetes deployment has not updated the Kubernetes cluster's API DNS entry to the correct IP address.  The API DNS IP address is the placeholder address that kops creates: 203.0.113.123.  Please wait about 5-10 minutes for a master to start, dns-controller to launch, and DNS to propagate.  The protokube container and dns-controller deployment logs may contain more diagnostic information.  Etcd and the API DNS entries must be updated for a kops Kubernetes cluster to start.

Validation Failed
W0729 14:04:51.008451    6342 validate_cluster.go:232] (will retry): cluster not yet healthy
INSTANCE GROUPS
NAME			ROLE	MACHINETYPE	MIN	MAX	SUBNETS
master-eu-west-3a	Master	c5.large	1	1	eu-west-3a
nodes-eu-west-3a	Node	t3.medium	4	4	eu-west-3a

NODE STATUS
NAME	ROLE	READY

VALIDATION ERRORS
KIND	NAME		MESSAGE
dns	apiserver	Validation Failed

The dns-controller Kubernetes deployment has not updated the Kubernetes cluster's API DNS entry to the correct IP address.  The API DNS IP address is the placeholder address that kops creates: 203.0.113.123.  Please wait about 5-10 minutes for a master to start, dns-controller to launch, and DNS to propagate.  The protokube container and dns-controller deployment logs may contain more diagnostic information.  Etcd and the API DNS entries must be updated for a kops Kubernetes cluster to start.

Validation Failed
W0729 14:05:01.095440    6342 validate_cluster.go:232] (will retry): cluster not yet healthy
INSTANCE GROUPS
NAME			ROLE	MACHINETYPE	MIN	MAX	SUBNETS
master-eu-west-3a	Master	c5.large	1	1	eu-west-3a
nodes-eu-west-3a	Node	t3.medium	4	4	eu-west-3a

NODE STATUS
NAME	ROLE	READY

VALIDATION ERRORS
KIND	NAME		MESSAGE
dns	apiserver	Validation Failed

The dns-controller Kubernetes deployment has not updated the Kubernetes cluster's API DNS entry to the correct IP address.  The API DNS IP address is the placeholder address that kops creates: 203.0.113.123.  Please wait about 5-10 minutes for a master to start, dns-controller to launch, and DNS to propagate.  The protokube container and dns-controller deployment logs may contain more diagnostic information.  Etcd and the API DNS entries must be updated for a kops Kubernetes cluster to start.

Validation Failed
W0729 14:05:11.130974    6342 validate_cluster.go:232] (will retry): cluster not yet healthy
INSTANCE GROUPS
NAME			ROLE	MACHINETYPE	MIN	MAX	SUBNETS
master-eu-west-3a	Master	c5.large	1	1	eu-west-3a
nodes-eu-west-3a	Node	t3.medium	4	4	eu-west-3a

NODE STATUS
NAME	ROLE	READY

VALIDATION ERRORS
KIND	NAME		MESSAGE
dns	apiserver	Validation Failed

The dns-controller Kubernetes deployment has not updated the Kubernetes cluster's API DNS entry to the correct IP address.  The API DNS IP address is the placeholder address that kops creates: 203.0.113.123.  Please wait about 5-10 minutes for a master to start, dns-controller to launch, and DNS to propagate.  The protokube container and dns-controller deployment logs may contain more diagnostic information.  Etcd and the API DNS entries must be updated for a kops Kubernetes cluster to start.

Validation Failed
W0729 14:05:21.170999    6342 validate_cluster.go:232] (will retry): cluster not yet healthy
INSTANCE GROUPS
NAME			ROLE	MACHINETYPE	MIN	MAX	SUBNETS
master-eu-west-3a	Master	c5.large	1	1	eu-west-3a
nodes-eu-west-3a	Node	t3.medium	4	4	eu-west-3a

NODE STATUS
NAME	ROLE	READY

VALIDATION ERRORS
KIND	NAME		MESSAGE
dns	apiserver	Validation Failed

The dns-controller Kubernetes deployment has not updated the Kubernetes cluster's API DNS entry to the correct IP address.  The API DNS IP address is the placeholder address that kops creates: 203.0.113.123.  Please wait about 5-10 minutes for a master to start, dns-controller to launch, and DNS to propagate.  The protokube container and dns-controller deployment logs may contain more diagnostic information.  Etcd and the API DNS entries must be updated for a kops Kubernetes cluster to start.

Validation Failed
W0729 14:05:31.220612    6342 validate_cluster.go:232] (will retry): cluster not yet healthy
INSTANCE GROUPS
NAME			ROLE	MACHINETYPE	MIN	MAX	SUBNETS
master-eu-west-3a	Master	c5.large	1	1	eu-west-3a
nodes-eu-west-3a	Node	t3.medium	4	4	eu-west-3a

NODE STATUS
NAME	ROLE	READY

VALIDATION ERRORS
KIND	NAME		MESSAGE
dns	apiserver	Validation Failed

The dns-controller Kubernetes deployment has not updated the Kubernetes cluster's API DNS entry to the correct IP address.  The API DNS IP address is the placeholder address that kops creates: 203.0.113.123.  Please wait about 5-10 minutes for a master to start, dns-controller to launch, and DNS to propagate.  The protokube container and dns-controller deployment logs may contain more diagnostic information.  Etcd and the API DNS entries must be updated for a kops Kubernetes cluster to start.

Validation Failed
W0729 14:05:41.251980    6342 validate_cluster.go:232] (will retry): cluster not yet healthy
INSTANCE GROUPS
NAME			ROLE	MACHINETYPE	MIN	MAX	SUBNETS
master-eu-west-3a	Master	c5.large	1	1	eu-west-3a
nodes-eu-west-3a	Node	t3.medium	4	4	eu-west-3a

NODE STATUS
NAME	ROLE	READY

VALIDATION ERRORS
KIND	NAME		MESSAGE
dns	apiserver	Validation Failed

The dns-controller Kubernetes deployment has not updated the Kubernetes cluster's API DNS entry to the correct IP address.  The API DNS IP address is the placeholder address that kops creates: 203.0.113.123.  Please wait about 5-10 minutes for a master to start, dns-controller to launch, and DNS to propagate.  The protokube container and dns-controller deployment logs may contain more diagnostic information.  Etcd and the API DNS entries must be updated for a kops Kubernetes cluster to start.

Validation Failed
W0729 14:05:51.302408    6342 validate_cluster.go:232] (will retry): cluster not yet healthy
INSTANCE GROUPS
NAME			ROLE	MACHINETYPE	MIN	MAX	SUBNETS
master-eu-west-3a	Master	c5.large	1	1	eu-west-3a
nodes-eu-west-3a	Node	t3.medium	4	4	eu-west-3a

NODE STATUS
NAME	ROLE	READY

VALIDATION ERRORS
KIND	NAME		MESSAGE
dns	apiserver	Validation Failed

The dns-controller Kubernetes deployment has not updated the Kubernetes cluster's API DNS entry to the correct IP address.  The API DNS IP address is the placeholder address that kops creates: 203.0.113.123.  Please wait about 5-10 minutes for a master to start, dns-controller to launch, and DNS to propagate.  The protokube container and dns-controller deployment logs may contain more diagnostic information.  Etcd and the API DNS entries must be updated for a kops Kubernetes cluster to start.

Validation Failed
W0729 14:06:01.351630    6342 validate_cluster.go:232] (will retry): cluster not yet healthy
INSTANCE GROUPS
NAME			ROLE	MACHINETYPE	MIN	MAX	SUBNETS
master-eu-west-3a	Master	c5.large	1	1	eu-west-3a
nodes-eu-west-3a	Node	t3.medium	4	4	eu-west-3a

NODE STATUS
NAME	ROLE	READY

VALIDATION ERRORS
KIND	NAME		MESSAGE
dns	apiserver	Validation Failed

The dns-controller Kubernetes deployment has not updated the Kubernetes cluster's API DNS entry to the correct IP address.  The API DNS IP address is the placeholder address that kops creates: 203.0.113.123.  Please wait about 5-10 minutes for a master to start, dns-controller to launch, and DNS to propagate.  The protokube container and dns-controller deployment logs may contain more diagnostic information.  Etcd and the API DNS entries must be updated for a kops Kubernetes cluster to start.

Validation Failed
W0729 14:06:11.402325    6342 validate_cluster.go:232] (will retry): cluster not yet healthy
INSTANCE GROUPS
NAME			ROLE	MACHINETYPE	MIN	MAX	SUBNETS
master-eu-west-3a	Master	c5.large	1	1	eu-west-3a
nodes-eu-west-3a	Node	t3.medium	4	4	eu-west-3a

NODE STATUS
NAME	ROLE	READY

VALIDATION ERRORS
KIND	NAME		MESSAGE
dns	apiserver	Validation Failed

The dns-controller Kubernetes deployment has not updated the Kubernetes cluster's API DNS entry to the correct IP address.  The API DNS IP address is the placeholder address that kops creates: 203.0.113.123.  Please wait about 5-10 minutes for a master to start, dns-controller to launch, and DNS to propagate.  The protokube container and dns-controller deployment logs may contain more diagnostic information.  Etcd and the API DNS entries must be updated for a kops Kubernetes cluster to start.

Validation Failed
W0729 14:06:21.442217    6342 validate_cluster.go:232] (will retry): cluster not yet healthy
INSTANCE GROUPS
NAME			ROLE	MACHINETYPE	MIN	MAX	SUBNETS
master-eu-west-3a	Master	c5.large	1	1	eu-west-3a
nodes-eu-west-3a	Node	t3.medium	4	4	eu-west-3a

NODE STATUS
NAME	ROLE	READY

VALIDATION ERRORS
KIND	NAME		MESSAGE
dns	apiserver	Validation Failed

The dns-controller Kubernetes deployment has not updated the Kubernetes cluster's API DNS entry to the correct IP address.  The API DNS IP address is the placeholder address that kops creates: 203.0.113.123.  Please wait about 5-10 minutes for a master to start, dns-controller to launch, and DNS to propagate.  The protokube container and dns-controller deployment logs may contain more diagnostic information.  Etcd and the API DNS entries must be updated for a kops Kubernetes cluster to start.

Validation Failed
W0729 14:06:31.479420    6342 validate_cluster.go:232] (will retry): cluster not yet healthy
INSTANCE GROUPS
NAME			ROLE	MACHINETYPE	MIN	MAX	SUBNETS
master-eu-west-3a	Master	c5.large	1	1	eu-west-3a
nodes-eu-west-3a	Node	t3.medium	4	4	eu-west-3a

NODE STATUS
NAME	ROLE	READY

VALIDATION ERRORS
KIND	NAME		MESSAGE
dns	apiserver	Validation Failed

The dns-controller Kubernetes deployment has not updated the Kubernetes cluster's API DNS entry to the correct IP address.  The API DNS IP address is the placeholder address that kops creates: 203.0.113.123.  Please wait about 5-10 minutes for a master to start, dns-controller to launch, and DNS to propagate.  The protokube container and dns-controller deployment logs may contain more diagnostic information.  Etcd and the API DNS entries must be updated for a kops Kubernetes cluster to start.

Validation Failed
W0729 14:06:41.521300    6342 validate_cluster.go:232] (will retry): cluster not yet healthy
INSTANCE GROUPS
NAME			ROLE	MACHINETYPE	MIN	MAX	SUBNETS
master-eu-west-3a	Master	c5.large	1	1	eu-west-3a
nodes-eu-west-3a	Node	t3.medium	4	4	eu-west-3a

NODE STATUS
NAME	ROLE	READY

VALIDATION ERRORS
KIND	NAME		MESSAGE
dns	apiserver	Validation Failed

The dns-controller Kubernetes deployment has not updated the Kubernetes cluster's API DNS entry to the correct IP address.  The API DNS IP address is the placeholder address that kops creates: 203.0.113.123.  Please wait about 5-10 minutes for a master to start, dns-controller to launch, and DNS to propagate.  The protokube container and dns-controller deployment logs may contain more diagnostic information.  Etcd and the API DNS entries must be updated for a kops Kubernetes cluster to start.

Validation Failed
W0729 14:06:51.556167    6342 validate_cluster.go:232] (will retry): cluster not yet healthy
INSTANCE GROUPS
NAME			ROLE	MACHINETYPE	MIN	MAX	SUBNETS
master-eu-west-3a	Master	c5.large	1	1	eu-west-3a
nodes-eu-west-3a	Node	t3.medium	4	4	eu-west-3a

NODE STATUS
NAME	ROLE	READY

VALIDATION ERRORS
KIND	NAME		MESSAGE
dns	apiserver	Validation Failed

The dns-controller Kubernetes deployment has not updated the Kubernetes cluster's API DNS entry to the correct IP address.  The API DNS IP address is the placeholder address that kops creates: 203.0.113.123.  Please wait about 5-10 minutes for a master to start, dns-controller to launch, and DNS to propagate.  The protokube container and dns-controller deployment logs may contain more diagnostic information.  Etcd and the API DNS entries must be updated for a kops Kubernetes cluster to start.

Validation Failed
W0729 14:07:01.592357    6342 validate_cluster.go:232] (will retry): cluster not yet healthy
INSTANCE GROUPS
NAME			ROLE	MACHINETYPE	MIN	MAX	SUBNETS
master-eu-west-3a	Master	c5.large	1	1	eu-west-3a
nodes-eu-west-3a	Node	t3.medium	4	4	eu-west-3a

NODE STATUS
NAME	ROLE	READY

VALIDATION ERRORS
KIND	NAME		MESSAGE
dns	apiserver	Validation Failed

The dns-controller Kubernetes deployment has not updated the Kubernetes cluster's API DNS entry to the correct IP address.  The API DNS IP address is the placeholder address that kops creates: 203.0.113.123.  Please wait about 5-10 minutes for a master to start, dns-controller to launch, and DNS to propagate.  The protokube container and dns-controller deployment logs may contain more diagnostic information.  Etcd and the API DNS entries must be updated for a kops Kubernetes cluster to start.

Validation Failed
W0729 14:07:11.643461    6342 validate_cluster.go:232] (will retry): cluster not yet healthy
INSTANCE GROUPS
NAME			ROLE	MACHINETYPE	MIN	MAX	SUBNETS
master-eu-west-3a	Master	c5.large	1	1	eu-west-3a
nodes-eu-west-3a	Node	t3.medium	4	4	eu-west-3a

... skipping 19 lines ...
Pod	kube-system/coredns-autoscaler-6d96c59bbf-n7tw9					system-cluster-critical pod "coredns-autoscaler-6d96c59bbf-n7tw9" is pending
Pod	kube-system/etcd-manager-main-ip-172-20-0-172.eu-west-3.compute.internal	system-cluster-critical pod "etcd-manager-main-ip-172-20-0-172.eu-west-3.compute.internal" is pending
Pod	kube-system/metrics-server-7c9d469d74-9mk8g					system-cluster-critical pod "metrics-server-7c9d469d74-9mk8g" is pending
Pod	kube-system/metrics-server-7c9d469d74-jxdmf					system-cluster-critical pod "metrics-server-7c9d469d74-jxdmf" is pending
Pod	kube-system/node-local-dns-zcnz9						system-node-critical pod "node-local-dns-zcnz9" is pending

Validation Failed
W0729 14:07:24.271133    6342 validate_cluster.go:232] (will retry): cluster not yet healthy
INSTANCE GROUPS
NAME			ROLE	MACHINETYPE	MIN	MAX	SUBNETS
master-eu-west-3a	Master	c5.large	1	1	eu-west-3a
nodes-eu-west-3a	Node	t3.medium	4	4	eu-west-3a

... skipping 26 lines ...
Pod	kube-system/kube-proxy-ip-172-20-0-30.eu-west-3.compute.internal	system-node-critical pod "kube-proxy-ip-172-20-0-30.eu-west-3.compute.internal" is pending
Pod	kube-system/metrics-server-7c9d469d74-9mk8g				system-cluster-critical pod "metrics-server-7c9d469d74-9mk8g" is pending
Pod	kube-system/metrics-server-7c9d469d74-jxdmf				system-cluster-critical pod "metrics-server-7c9d469d74-jxdmf" is pending
Pod	kube-system/node-local-dns-cmk9r					system-node-critical pod "node-local-dns-cmk9r" is pending
Pod	kube-system/node-local-dns-dd9dj					system-node-critical pod "node-local-dns-dd9dj" is pending

Validation Failed
W0729 14:07:36.346511    6342 validate_cluster.go:232] (will retry): cluster not yet healthy
INSTANCE GROUPS
NAME			ROLE	MACHINETYPE	MIN	MAX	SUBNETS
master-eu-west-3a	Master	c5.large	1	1	eu-west-3a
nodes-eu-west-3a	Node	t3.medium	4	4	eu-west-3a

... skipping 24 lines ...
Pod	kube-system/coredns-78cd66cbc9-d7trt				system-cluster-critical pod "coredns-78cd66cbc9-d7trt" is pending
Pod	kube-system/coredns-autoscaler-6d96c59bbf-n7tw9			system-cluster-critical pod "coredns-autoscaler-6d96c59bbf-n7tw9" is pending
Pod	kube-system/metrics-server-7c9d469d74-9mk8g			system-cluster-critical pod "metrics-server-7c9d469d74-9mk8g" is pending
Pod	kube-system/metrics-server-7c9d469d74-jxdmf			system-cluster-critical pod "metrics-server-7c9d469d74-jxdmf" is pending
Pod	kube-system/node-local-dns-dd9dj				system-node-critical pod "node-local-dns-dd9dj" is pending

Validation Failed
W0729 14:07:48.310292    6342 validate_cluster.go:232] (will retry): cluster not yet healthy
INSTANCE GROUPS
NAME			ROLE	MACHINETYPE	MIN	MAX	SUBNETS
master-eu-west-3a	Master	c5.large	1	1	eu-west-3a
nodes-eu-west-3a	Node	t3.medium	4	4	eu-west-3a

... skipping 16 lines ...
Pod	kube-system/cilium-d5st4				system-node-critical pod "cilium-d5st4" is not ready (cilium-agent)
Pod	kube-system/cilium-j2hdj				system-node-critical pod "cilium-j2hdj" is not ready (cilium-agent)
Pod	kube-system/coredns-78cd66cbc9-8h8gt			system-cluster-critical pod "coredns-78cd66cbc9-8h8gt" is pending
Pod	kube-system/metrics-server-7c9d469d74-9mk8g		system-cluster-critical pod "metrics-server-7c9d469d74-9mk8g" is not ready (metrics-server)
Pod	kube-system/metrics-server-7c9d469d74-jxdmf		system-cluster-critical pod "metrics-server-7c9d469d74-jxdmf" is pending

Validation Failed
W0729 14:08:00.189065    6342 validate_cluster.go:232] (will retry): cluster not yet healthy
INSTANCE GROUPS
NAME			ROLE	MACHINETYPE	MIN	MAX	SUBNETS
master-eu-west-3a	Master	c5.large	1	1	eu-west-3a
nodes-eu-west-3a	Node	t3.medium	4	4	eu-west-3a

... skipping 10 lines ...
Pod	kube-system/aws-load-balancer-controller-c6d5b7dd-dx8xk	system-cluster-critical pod "aws-load-balancer-controller-c6d5b7dd-dx8xk" is pending
Pod	kube-system/cilium-4hlj9				system-node-critical pod "cilium-4hlj9" is not ready (cilium-agent)
Pod	kube-system/cilium-6lmsk				system-node-critical pod "cilium-6lmsk" is not ready (cilium-agent)
Pod	kube-system/metrics-server-7c9d469d74-9mk8g		system-cluster-critical pod "metrics-server-7c9d469d74-9mk8g" is not ready (metrics-server)
Pod	kube-system/metrics-server-7c9d469d74-jxdmf		system-cluster-critical pod "metrics-server-7c9d469d74-jxdmf" is not ready (metrics-server)

Validation Failed
W0729 14:08:12.100895    6342 validate_cluster.go:232] (will retry): cluster not yet healthy
INSTANCE GROUPS
NAME			ROLE	MACHINETYPE	MIN	MAX	SUBNETS
master-eu-west-3a	Master	c5.large	1	1	eu-west-3a
nodes-eu-west-3a	Node	t3.medium	4	4	eu-west-3a

... skipping 7 lines ...

VALIDATION ERRORS
KIND	NAME							MESSAGE
Pod	kube-system/aws-load-balancer-controller-c6d5b7dd-dx8xk	system-cluster-critical pod "aws-load-balancer-controller-c6d5b7dd-dx8xk" is pending
Pod	kube-system/metrics-server-7c9d469d74-jxdmf		system-cluster-critical pod "metrics-server-7c9d469d74-jxdmf" is not ready (metrics-server)

Validation Failed
W0729 14:08:24.134761    6342 validate_cluster.go:232] (will retry): cluster not yet healthy
INSTANCE GROUPS
NAME			ROLE	MACHINETYPE	MIN	MAX	SUBNETS
master-eu-west-3a	Master	c5.large	1	1	eu-west-3a
nodes-eu-west-3a	Node	t3.medium	4	4	eu-west-3a

... skipping 6 lines ...
ip-172-20-0-51.eu-west-3.compute.internal	node	True

VALIDATION ERRORS
KIND	NAME							MESSAGE
Pod	kube-system/aws-load-balancer-controller-c6d5b7dd-dx8xk	system-cluster-critical pod "aws-load-balancer-controller-c6d5b7dd-dx8xk" is pending

Validation Failed
W0729 14:08:36.099004    6342 validate_cluster.go:232] (will retry): cluster not yet healthy
INSTANCE GROUPS
NAME			ROLE	MACHINETYPE	MIN	MAX	SUBNETS
master-eu-west-3a	Master	c5.large	1	1	eu-west-3a
nodes-eu-west-3a	Node	t3.medium	4	4	eu-west-3a

... skipping 6 lines ...
ip-172-20-0-51.eu-west-3.compute.internal	node	True

VALIDATION ERRORS
KIND	NAME							MESSAGE
Pod	kube-system/aws-load-balancer-controller-c6d5b7dd-dx8xk	system-cluster-critical pod "aws-load-balancer-controller-c6d5b7dd-dx8xk" is pending

Validation Failed
W0729 14:08:48.086067    6342 validate_cluster.go:232] (will retry): cluster not yet healthy
INSTANCE GROUPS
NAME			ROLE	MACHINETYPE	MIN	MAX	SUBNETS
master-eu-west-3a	Master	c5.large	1	1	eu-west-3a
nodes-eu-west-3a	Node	t3.medium	4	4	eu-west-3a

... skipping 6 lines ...
ip-172-20-0-51.eu-west-3.compute.internal	node	True

VALIDATION ERRORS
KIND	NAME							MESSAGE
Pod	kube-system/aws-load-balancer-controller-c6d5b7dd-dx8xk	system-cluster-critical pod "aws-load-balancer-controller-c6d5b7dd-dx8xk" is pending

Validation Failed
W0729 14:09:00.127022    6342 validate_cluster.go:232] (will retry): cluster not yet healthy
INSTANCE GROUPS
NAME			ROLE	MACHINETYPE	MIN	MAX	SUBNETS
master-eu-west-3a	Master	c5.large	1	1	eu-west-3a
nodes-eu-west-3a	Node	t3.medium	4	4	eu-west-3a

... skipping 6 lines ...
ip-172-20-0-51.eu-west-3.compute.internal	node	True

VALIDATION ERRORS
KIND	NAME							MESSAGE
Pod	kube-system/aws-load-balancer-controller-c6d5b7dd-dx8xk	system-cluster-critical pod "aws-load-balancer-controller-c6d5b7dd-dx8xk" is pending

Validation Failed
W0729 14:09:12.118815    6342 validate_cluster.go:232] (will retry): cluster not yet healthy
INSTANCE GROUPS
NAME			ROLE	MACHINETYPE	MIN	MAX	SUBNETS
master-eu-west-3a	Master	c5.large	1	1	eu-west-3a
nodes-eu-west-3a	Node	t3.medium	4	4	eu-west-3a

... skipping 6 lines ...
ip-172-20-0-51.eu-west-3.compute.internal	node	True

VALIDATION ERRORS
KIND	NAME							MESSAGE
Pod	kube-system/aws-load-balancer-controller-c6d5b7dd-dx8xk	system-cluster-critical pod "aws-load-balancer-controller-c6d5b7dd-dx8xk" is pending

Validation Failed
W0729 14:09:23.978596    6342 validate_cluster.go:232] (will retry): cluster not yet healthy
INSTANCE GROUPS
NAME			ROLE	MACHINETYPE	MIN	MAX	SUBNETS
master-eu-west-3a	Master	c5.large	1	1	eu-west-3a
nodes-eu-west-3a	Node	t3.medium	4	4	eu-west-3a

... skipping 1027 lines ...
I0729 14:13:58.696970    6435 channel_version.go:140] manifest Match for "dns-controller.addons.k8s.io": Channel=s3://k8s-kops-prow/e2e-414ec24bfc-41f44.test-cncf-aws.k8s.io/addons/bootstrap-channel.yaml Id=k8s-1.12 ManifestHash=d6a10da0267cdaec22667d76cffe6d9ccb0bceda4ce613269d71b439ecfba213 SystemGeneration=1
NAME			CURRENT	UPDATE									PKI
networking.cilium.io	-	0d7e665f356f97c0137cda33453dbd5d40d5fbca30b62df8ef30e2db894a36a1	no
I0729 14:13:58.905794    6435 addon.go:188] Applying update from "s3://k8s-kops-prow/e2e-414ec24bfc-41f44.test-cncf-aws.k8s.io/addons/networking.cilium.io/k8s-1.16-v1.11.yaml"
I0729 14:13:58.905854    6435 s3fs.go:329] Reading file "s3://k8s-kops-prow/e2e-414ec24bfc-41f44.test-cncf-aws.k8s.io/addons/networking.cilium.io/k8s-1.16-v1.11.yaml"
I0729 14:13:59.013943    6435 apply.go:83] Running command: kubectl apply -f /tmp/channel2186814885/manifest.yaml --server-side --force-conflicts --field-manager=kops
I0729 14:14:01.877541    6435 apply.go:86] error running kubectl apply -f /tmp/channel2186814885/manifest.yaml --server-side --force-conflicts --field-manager=kops
I0729 14:14:01.877582    6435 apply.go:87] serviceaccount/cilium serverside-applied
serviceaccount/cilium-operator serverside-applied
serviceaccount/hubble-relay serverside-applied
configmap/cilium-config serverside-applied
configmap/hubble-relay-config serverside-applied
clusterrole.rbac.authorization.k8s.io/cilium serverside-applied
... skipping 5 lines ...
service/hubble-relay serverside-applied
daemonset.apps/cilium serverside-applied
deployment.apps/cilium-operator serverside-applied
certificate.cert-manager.io/hubble-server-certs serverside-applied
certificate.cert-manager.io/hubble-relay-client-certs serverside-applied
poddisruptionbudget.policy/cilium-operator serverside-applied
Error from server: failed to create typed patch object: .spec.template.spec.containers[name="hubble-relay"].ports: element 0: associative list with keys has an element that omits key field "protocol" (and doesn't have default value)
E0729 14:14:01.877603    6435 apply.go:52] failed to apply the manifest: error running kubectl: exit status 1
I0729 14:14:01.877684    6435 apply.go:83] Running command: kubectl replace -f /tmp/channel2186814885/manifest.yaml --field-manager=kops
I0729 14:14:06.197516    6435 apply.go:86] error running kubectl replace -f /tmp/channel2186814885/manifest.yaml --field-manager=kops
I0729 14:14:06.197555    6435 apply.go:87] serviceaccount/cilium replaced
serviceaccount/cilium-operator replaced
serviceaccount/hubble-relay replaced
configmap/cilium-config replaced
configmap/hubble-relay-config replaced
clusterrole.rbac.authorization.k8s.io/cilium replaced
... skipping 4 lines ...
clusterrolebinding.rbac.authorization.k8s.io/hubble-relay replaced
daemonset.apps/cilium replaced
deployment.apps/cilium-operator replaced
certificate.cert-manager.io/hubble-server-certs replaced
certificate.cert-manager.io/hubble-relay-client-certs replaced
poddisruptionbudget.policy/cilium-operator replaced
Error from server (Invalid): error when replacing "/tmp/channel2186814885/manifest.yaml": Service "hubble-relay" is invalid: spec.clusterIP: Invalid value: "": field is immutable
Error from server (NotFound): error when replacing "/tmp/channel2186814885/manifest.yaml": deployments.apps "hubble-relay" not found
E0729 14:14:06.197585    6435 apply.go:61] failed to replace manifest: error running kubectl: exit status 1
I0729 14:14:06.197678    6435 apply.go:83] Running command: kubectl apply -f /tmp/channel2186814885/manifest.yaml --server-side --force-conflicts --field-manager=kops
I0729 14:14:08.782310    6435 apply.go:86] error running kubectl apply -f /tmp/channel2186814885/manifest.yaml --server-side --force-conflicts --field-manager=kops
I0729 14:14:08.782346    6435 apply.go:87] serviceaccount/cilium serverside-applied
serviceaccount/cilium-operator serverside-applied
serviceaccount/hubble-relay serverside-applied
configmap/cilium-config serverside-applied
configmap/hubble-relay-config serverside-applied
clusterrole.rbac.authorization.k8s.io/cilium serverside-applied
... skipping 5 lines ...
service/hubble-relay serverside-applied
daemonset.apps/cilium serverside-applied
deployment.apps/cilium-operator serverside-applied
certificate.cert-manager.io/hubble-server-certs serverside-applied
certificate.cert-manager.io/hubble-relay-client-certs serverside-applied
poddisruptionbudget.policy/cilium-operator serverside-applied
Error from server: failed to create typed patch object: .spec.template.spec.containers[name="hubble-relay"].ports: element 0: associative list with keys has an element that omits key field "protocol" (and doesn't have default value)

updating "networking.cilium.io": error applying update from "s3://k8s-kops-prow/e2e-414ec24bfc-41f44.test-cncf-aws.k8s.io/addons/networking.cilium.io/k8s-1.16-v1.11.yaml": failed to apply the manifest: error running kubectl: exit status 1
+ kops-finish
+ kubetest2 kops -v=2 --cloud-provider=aws --cluster-name=e2e-414ec24bfc-41f44.test-cncf-aws.k8s.io --kops-root=/home/prow/go/src/k8s.io/kops --admin-access= --env=KOPS_FEATURE_FLAGS=SpecOverrideFlag --kops-binary-path=/tmp/kops.Gc2Stcgpy --down
I0729 14:14:08.809759    6483 featureflag.go:162] FeatureFlag "SpecOverrideFlag"=true
I0729 14:14:08.810808    6483 app.go:62] The files in RunDir shall not be part of Artifacts
I0729 14:14:08.810830    6483 app.go:63] pass rundir-in-artifacts flag True for RunDir to be part of Artifacts
I0729 14:14:08.810859    6483 app.go:65] RunDir for this run: "/home/prow/go/src/k8s.io/kops/_rundir/d1d52443-0f46-11ed-b6e4-d28c295d3c4e"
... skipping 292 lines ...