This job view page is being replaced by Spyglass soon. Check out the new job view.
ResultFAILURE
Tests 0 failed / 1 succeeded
Started2022-07-29 08:01
Elapsed19m2s
Revisionmaster

No Test Failures!


Show 1 Passed Tests

Error lines from build-log.txt

... skipping 212 lines ...
I0729 08:02:20.487366    6154 app.go:63] pass rundir-in-artifacts flag True for RunDir to be part of Artifacts
I0729 08:02:20.487412    6154 app.go:65] RunDir for this run: "/home/prow/go/src/k8s.io/kops/_rundir/871c3a8b-0f14-11ed-b6e4-d28c295d3c4e"
I0729 08:02:20.538710    6154 app.go:129] ID for this run: "871c3a8b-0f14-11ed-b6e4-d28c295d3c4e"
I0729 08:02:20.538980    6154 local.go:42] ⚙️ ssh-keygen -t ed25519 -N  -q -f /tmp/kops/e2e-414ec24bfc-41f44.test-cncf-aws.k8s.io/id_ed25519
I0729 08:02:20.545588    6154 dumplogs.go:45] /tmp/kops.EvnerqtlZ toolbox dump --name e2e-414ec24bfc-41f44.test-cncf-aws.k8s.io --dir /logs/artifacts --private-key /tmp/kops/e2e-414ec24bfc-41f44.test-cncf-aws.k8s.io/id_ed25519 --ssh-user ubuntu
I0729 08:02:20.545632    6154 local.go:42] ⚙️ /tmp/kops.EvnerqtlZ toolbox dump --name e2e-414ec24bfc-41f44.test-cncf-aws.k8s.io --dir /logs/artifacts --private-key /tmp/kops/e2e-414ec24bfc-41f44.test-cncf-aws.k8s.io/id_ed25519 --ssh-user ubuntu
W0729 08:02:21.037861    6154 down.go:34] Dumping cluster logs at the start of Down() failed: exit status 1
I0729 08:02:21.037902    6154 down.go:48] /tmp/kops.EvnerqtlZ delete cluster --name e2e-414ec24bfc-41f44.test-cncf-aws.k8s.io --yes
I0729 08:02:21.037913    6154 local.go:42] ⚙️ /tmp/kops.EvnerqtlZ delete cluster --name e2e-414ec24bfc-41f44.test-cncf-aws.k8s.io --yes
I0729 08:02:21.059752    6175 featureflag.go:162] FeatureFlag "SpecOverrideFlag"=true
I0729 08:02:21.059846    6175 featureflag.go:162] FeatureFlag "SpecOverrideFlag"=true
Error: error reading cluster configuration: Cluster.kops.k8s.io "e2e-414ec24bfc-41f44.test-cncf-aws.k8s.io" not found
Error: exit status 1
+ echo 'kubetest2 down failed'
kubetest2 down failed
+ [[ v == \v ]]
+ KOPS_BASE_URL=
++ kops-download-release v1.24.0
++ local kops
+++ mktemp -t kops.XXXXXXXXX
++ kops=/tmp/kops.nFogXSZFo
... skipping 10 lines ...
I0729 08:02:25.212334    6209 app.go:63] pass rundir-in-artifacts flag True for RunDir to be part of Artifacts
I0729 08:02:25.212362    6209 app.go:65] RunDir for this run: "/home/prow/go/src/k8s.io/kops/_rundir/871c3a8b-0f14-11ed-b6e4-d28c295d3c4e"
I0729 08:02:25.217698    6209 app.go:129] ID for this run: "871c3a8b-0f14-11ed-b6e4-d28c295d3c4e"
I0729 08:02:25.217787    6209 up.go:44] Cleaning up any leaked resources from previous cluster
I0729 08:02:25.217815    6209 dumplogs.go:45] /tmp/kops.nFogXSZFo toolbox dump --name e2e-414ec24bfc-41f44.test-cncf-aws.k8s.io --dir /logs/artifacts --private-key /tmp/kops/e2e-414ec24bfc-41f44.test-cncf-aws.k8s.io/id_ed25519 --ssh-user ubuntu
I0729 08:02:25.217840    6209 local.go:42] ⚙️ /tmp/kops.nFogXSZFo toolbox dump --name e2e-414ec24bfc-41f44.test-cncf-aws.k8s.io --dir /logs/artifacts --private-key /tmp/kops/e2e-414ec24bfc-41f44.test-cncf-aws.k8s.io/id_ed25519 --ssh-user ubuntu
W0729 08:02:25.712380    6209 down.go:34] Dumping cluster logs at the start of Down() failed: exit status 1
I0729 08:02:25.712543    6209 down.go:48] /tmp/kops.nFogXSZFo delete cluster --name e2e-414ec24bfc-41f44.test-cncf-aws.k8s.io --yes
I0729 08:02:25.712574    6209 local.go:42] ⚙️ /tmp/kops.nFogXSZFo delete cluster --name e2e-414ec24bfc-41f44.test-cncf-aws.k8s.io --yes
I0729 08:02:25.732715    6229 featureflag.go:162] FeatureFlag "SpecOverrideFlag"=true
I0729 08:02:25.732805    6229 featureflag.go:162] FeatureFlag "SpecOverrideFlag"=true
Error: error reading cluster configuration: Cluster.kops.k8s.io "e2e-414ec24bfc-41f44.test-cncf-aws.k8s.io" not found
I0729 08:02:26.181417    6209 http.go:37] curl http://metadata.google.internal/computeMetadata/v1/instance/network-interfaces/0/access-configs/0/external-ip
2022/07/29 08:02:26 failed to get external ip from metadata service: http://metadata.google.internal/computeMetadata/v1/instance/network-interfaces/0/access-configs/0/external-ip returned 404
I0729 08:02:26.196154    6209 http.go:37] curl https://ip.jsb.workers.dev
I0729 08:02:26.306354    6209 template.go:58] /tmp/kops.nFogXSZFo toolbox template --template tests/e2e/templates/many-addons.yaml.tmpl --output /tmp/kops-template2939000803/manifest.yaml --values /tmp/kops-template2939000803/values.yaml --name e2e-414ec24bfc-41f44.test-cncf-aws.k8s.io
I0729 08:02:26.306404    6209 local.go:42] ⚙️ /tmp/kops.nFogXSZFo toolbox template --template tests/e2e/templates/many-addons.yaml.tmpl --output /tmp/kops-template2939000803/manifest.yaml --values /tmp/kops-template2939000803/values.yaml --name e2e-414ec24bfc-41f44.test-cncf-aws.k8s.io
I0729 08:02:26.328201    6242 featureflag.go:162] FeatureFlag "SpecOverrideFlag"=true
I0729 08:02:26.328299    6242 featureflag.go:162] FeatureFlag "SpecOverrideFlag"=true
I0729 08:02:26.467848    6209 create.go:33] /tmp/kops.nFogXSZFo create --filename /tmp/kops-template2939000803/manifest.yaml --name e2e-414ec24bfc-41f44.test-cncf-aws.k8s.io
... skipping 54 lines ...
I0729 08:03:03.635202    6209 up.go:243] /tmp/kops.nFogXSZFo validate cluster --name e2e-414ec24bfc-41f44.test-cncf-aws.k8s.io --count 10 --wait 15m0s
I0729 08:03:03.635270    6209 local.go:42] ⚙️ /tmp/kops.nFogXSZFo validate cluster --name e2e-414ec24bfc-41f44.test-cncf-aws.k8s.io --count 10 --wait 15m0s
I0729 08:03:03.657356    6283 featureflag.go:162] FeatureFlag "SpecOverrideFlag"=true
I0729 08:03:03.657473    6283 featureflag.go:162] FeatureFlag "SpecOverrideFlag"=true
Validating cluster e2e-414ec24bfc-41f44.test-cncf-aws.k8s.io

W0729 08:03:05.503449    6283 validate_cluster.go:184] (will retry): unexpected error during validation: unable to resolve Kubernetes cluster API URL dns: lookup api.e2e-414ec24bfc-41f44.test-cncf-aws.k8s.io on 10.63.240.10:53: no such host
INSTANCE GROUPS
NAME			ROLE	MACHINETYPE	MIN	MAX	SUBNETS
master-ap-south-1a	Master	c5.large	1	1	ap-south-1a
nodes-ap-south-1a	Node	t3.medium	4	4	ap-south-1a

NODE STATUS
NAME	ROLE	READY

VALIDATION ERRORS
KIND	NAME		MESSAGE
dns	apiserver	Validation Failed

The dns-controller Kubernetes deployment has not updated the Kubernetes cluster's API DNS entry to the correct IP address.  The API DNS IP address is the placeholder address that kops creates: 203.0.113.123.  Please wait about 5-10 minutes for a master to start, dns-controller to launch, and DNS to propagate.  The protokube container and dns-controller deployment logs may contain more diagnostic information.  Etcd and the API DNS entries must be updated for a kops Kubernetes cluster to start.

Validation Failed
W0729 08:03:15.542679    6283 validate_cluster.go:232] (will retry): cluster not yet healthy
INSTANCE GROUPS
NAME			ROLE	MACHINETYPE	MIN	MAX	SUBNETS
master-ap-south-1a	Master	c5.large	1	1	ap-south-1a
nodes-ap-south-1a	Node	t3.medium	4	4	ap-south-1a

NODE STATUS
NAME	ROLE	READY

VALIDATION ERRORS
KIND	NAME		MESSAGE
dns	apiserver	Validation Failed

The dns-controller Kubernetes deployment has not updated the Kubernetes cluster's API DNS entry to the correct IP address.  The API DNS IP address is the placeholder address that kops creates: 203.0.113.123.  Please wait about 5-10 minutes for a master to start, dns-controller to launch, and DNS to propagate.  The protokube container and dns-controller deployment logs may contain more diagnostic information.  Etcd and the API DNS entries must be updated for a kops Kubernetes cluster to start.

Validation Failed
W0729 08:03:25.580781    6283 validate_cluster.go:232] (will retry): cluster not yet healthy
INSTANCE GROUPS
NAME			ROLE	MACHINETYPE	MIN	MAX	SUBNETS
master-ap-south-1a	Master	c5.large	1	1	ap-south-1a
nodes-ap-south-1a	Node	t3.medium	4	4	ap-south-1a

NODE STATUS
NAME	ROLE	READY

VALIDATION ERRORS
KIND	NAME		MESSAGE
dns	apiserver	Validation Failed

The dns-controller Kubernetes deployment has not updated the Kubernetes cluster's API DNS entry to the correct IP address.  The API DNS IP address is the placeholder address that kops creates: 203.0.113.123.  Please wait about 5-10 minutes for a master to start, dns-controller to launch, and DNS to propagate.  The protokube container and dns-controller deployment logs may contain more diagnostic information.  Etcd and the API DNS entries must be updated for a kops Kubernetes cluster to start.

Validation Failed
W0729 08:03:35.618319    6283 validate_cluster.go:232] (will retry): cluster not yet healthy
INSTANCE GROUPS
NAME			ROLE	MACHINETYPE	MIN	MAX	SUBNETS
master-ap-south-1a	Master	c5.large	1	1	ap-south-1a
nodes-ap-south-1a	Node	t3.medium	4	4	ap-south-1a

NODE STATUS
NAME	ROLE	READY

VALIDATION ERRORS
KIND	NAME		MESSAGE
dns	apiserver	Validation Failed

The dns-controller Kubernetes deployment has not updated the Kubernetes cluster's API DNS entry to the correct IP address.  The API DNS IP address is the placeholder address that kops creates: 203.0.113.123.  Please wait about 5-10 minutes for a master to start, dns-controller to launch, and DNS to propagate.  The protokube container and dns-controller deployment logs may contain more diagnostic information.  Etcd and the API DNS entries must be updated for a kops Kubernetes cluster to start.

Validation Failed
W0729 08:03:45.668043    6283 validate_cluster.go:232] (will retry): cluster not yet healthy
INSTANCE GROUPS
NAME			ROLE	MACHINETYPE	MIN	MAX	SUBNETS
master-ap-south-1a	Master	c5.large	1	1	ap-south-1a
nodes-ap-south-1a	Node	t3.medium	4	4	ap-south-1a

NODE STATUS
NAME	ROLE	READY

VALIDATION ERRORS
KIND	NAME		MESSAGE
dns	apiserver	Validation Failed

The dns-controller Kubernetes deployment has not updated the Kubernetes cluster's API DNS entry to the correct IP address.  The API DNS IP address is the placeholder address that kops creates: 203.0.113.123.  Please wait about 5-10 minutes for a master to start, dns-controller to launch, and DNS to propagate.  The protokube container and dns-controller deployment logs may contain more diagnostic information.  Etcd and the API DNS entries must be updated for a kops Kubernetes cluster to start.

Validation Failed
W0729 08:03:55.704683    6283 validate_cluster.go:232] (will retry): cluster not yet healthy
INSTANCE GROUPS
NAME			ROLE	MACHINETYPE	MIN	MAX	SUBNETS
master-ap-south-1a	Master	c5.large	1	1	ap-south-1a
nodes-ap-south-1a	Node	t3.medium	4	4	ap-south-1a

NODE STATUS
NAME	ROLE	READY

VALIDATION ERRORS
KIND	NAME		MESSAGE
dns	apiserver	Validation Failed

The dns-controller Kubernetes deployment has not updated the Kubernetes cluster's API DNS entry to the correct IP address.  The API DNS IP address is the placeholder address that kops creates: 203.0.113.123.  Please wait about 5-10 minutes for a master to start, dns-controller to launch, and DNS to propagate.  The protokube container and dns-controller deployment logs may contain more diagnostic information.  Etcd and the API DNS entries must be updated for a kops Kubernetes cluster to start.

Validation Failed
W0729 08:04:05.742236    6283 validate_cluster.go:232] (will retry): cluster not yet healthy
INSTANCE GROUPS
NAME			ROLE	MACHINETYPE	MIN	MAX	SUBNETS
master-ap-south-1a	Master	c5.large	1	1	ap-south-1a
nodes-ap-south-1a	Node	t3.medium	4	4	ap-south-1a

NODE STATUS
NAME	ROLE	READY

VALIDATION ERRORS
KIND	NAME		MESSAGE
dns	apiserver	Validation Failed

The dns-controller Kubernetes deployment has not updated the Kubernetes cluster's API DNS entry to the correct IP address.  The API DNS IP address is the placeholder address that kops creates: 203.0.113.123.  Please wait about 5-10 minutes for a master to start, dns-controller to launch, and DNS to propagate.  The protokube container and dns-controller deployment logs may contain more diagnostic information.  Etcd and the API DNS entries must be updated for a kops Kubernetes cluster to start.

Validation Failed
W0729 08:04:15.791364    6283 validate_cluster.go:232] (will retry): cluster not yet healthy
INSTANCE GROUPS
NAME			ROLE	MACHINETYPE	MIN	MAX	SUBNETS
master-ap-south-1a	Master	c5.large	1	1	ap-south-1a
nodes-ap-south-1a	Node	t3.medium	4	4	ap-south-1a

NODE STATUS
NAME	ROLE	READY

VALIDATION ERRORS
KIND	NAME		MESSAGE
dns	apiserver	Validation Failed

The dns-controller Kubernetes deployment has not updated the Kubernetes cluster's API DNS entry to the correct IP address.  The API DNS IP address is the placeholder address that kops creates: 203.0.113.123.  Please wait about 5-10 minutes for a master to start, dns-controller to launch, and DNS to propagate.  The protokube container and dns-controller deployment logs may contain more diagnostic information.  Etcd and the API DNS entries must be updated for a kops Kubernetes cluster to start.

Validation Failed
W0729 08:04:25.828446    6283 validate_cluster.go:232] (will retry): cluster not yet healthy
INSTANCE GROUPS
NAME			ROLE	MACHINETYPE	MIN	MAX	SUBNETS
master-ap-south-1a	Master	c5.large	1	1	ap-south-1a
nodes-ap-south-1a	Node	t3.medium	4	4	ap-south-1a

NODE STATUS
NAME	ROLE	READY

VALIDATION ERRORS
KIND	NAME		MESSAGE
dns	apiserver	Validation Failed

The dns-controller Kubernetes deployment has not updated the Kubernetes cluster's API DNS entry to the correct IP address.  The API DNS IP address is the placeholder address that kops creates: 203.0.113.123.  Please wait about 5-10 minutes for a master to start, dns-controller to launch, and DNS to propagate.  The protokube container and dns-controller deployment logs may contain more diagnostic information.  Etcd and the API DNS entries must be updated for a kops Kubernetes cluster to start.

Validation Failed
W0729 08:04:35.867099    6283 validate_cluster.go:232] (will retry): cluster not yet healthy
INSTANCE GROUPS
NAME			ROLE	MACHINETYPE	MIN	MAX	SUBNETS
master-ap-south-1a	Master	c5.large	1	1	ap-south-1a
nodes-ap-south-1a	Node	t3.medium	4	4	ap-south-1a

NODE STATUS
NAME	ROLE	READY

VALIDATION ERRORS
KIND	NAME		MESSAGE
dns	apiserver	Validation Failed

The dns-controller Kubernetes deployment has not updated the Kubernetes cluster's API DNS entry to the correct IP address.  The API DNS IP address is the placeholder address that kops creates: 203.0.113.123.  Please wait about 5-10 minutes for a master to start, dns-controller to launch, and DNS to propagate.  The protokube container and dns-controller deployment logs may contain more diagnostic information.  Etcd and the API DNS entries must be updated for a kops Kubernetes cluster to start.

Validation Failed
W0729 08:04:45.907468    6283 validate_cluster.go:232] (will retry): cluster not yet healthy
INSTANCE GROUPS
NAME			ROLE	MACHINETYPE	MIN	MAX	SUBNETS
master-ap-south-1a	Master	c5.large	1	1	ap-south-1a
nodes-ap-south-1a	Node	t3.medium	4	4	ap-south-1a

NODE STATUS
NAME	ROLE	READY

VALIDATION ERRORS
KIND	NAME		MESSAGE
dns	apiserver	Validation Failed

The dns-controller Kubernetes deployment has not updated the Kubernetes cluster's API DNS entry to the correct IP address.  The API DNS IP address is the placeholder address that kops creates: 203.0.113.123.  Please wait about 5-10 minutes for a master to start, dns-controller to launch, and DNS to propagate.  The protokube container and dns-controller deployment logs may contain more diagnostic information.  Etcd and the API DNS entries must be updated for a kops Kubernetes cluster to start.

Validation Failed
W0729 08:04:55.946612    6283 validate_cluster.go:232] (will retry): cluster not yet healthy
INSTANCE GROUPS
NAME			ROLE	MACHINETYPE	MIN	MAX	SUBNETS
master-ap-south-1a	Master	c5.large	1	1	ap-south-1a
nodes-ap-south-1a	Node	t3.medium	4	4	ap-south-1a

NODE STATUS
NAME	ROLE	READY

VALIDATION ERRORS
KIND	NAME		MESSAGE
dns	apiserver	Validation Failed

The dns-controller Kubernetes deployment has not updated the Kubernetes cluster's API DNS entry to the correct IP address.  The API DNS IP address is the placeholder address that kops creates: 203.0.113.123.  Please wait about 5-10 minutes for a master to start, dns-controller to launch, and DNS to propagate.  The protokube container and dns-controller deployment logs may contain more diagnostic information.  Etcd and the API DNS entries must be updated for a kops Kubernetes cluster to start.

Validation Failed
W0729 08:05:05.997750    6283 validate_cluster.go:232] (will retry): cluster not yet healthy
INSTANCE GROUPS
NAME			ROLE	MACHINETYPE	MIN	MAX	SUBNETS
master-ap-south-1a	Master	c5.large	1	1	ap-south-1a
nodes-ap-south-1a	Node	t3.medium	4	4	ap-south-1a

NODE STATUS
NAME	ROLE	READY

VALIDATION ERRORS
KIND	NAME		MESSAGE
dns	apiserver	Validation Failed

The dns-controller Kubernetes deployment has not updated the Kubernetes cluster's API DNS entry to the correct IP address.  The API DNS IP address is the placeholder address that kops creates: 203.0.113.123.  Please wait about 5-10 minutes for a master to start, dns-controller to launch, and DNS to propagate.  The protokube container and dns-controller deployment logs may contain more diagnostic information.  Etcd and the API DNS entries must be updated for a kops Kubernetes cluster to start.

Validation Failed
W0729 08:05:16.050265    6283 validate_cluster.go:232] (will retry): cluster not yet healthy
INSTANCE GROUPS
NAME			ROLE	MACHINETYPE	MIN	MAX	SUBNETS
master-ap-south-1a	Master	c5.large	1	1	ap-south-1a
nodes-ap-south-1a	Node	t3.medium	4	4	ap-south-1a

NODE STATUS
NAME	ROLE	READY

VALIDATION ERRORS
KIND	NAME		MESSAGE
dns	apiserver	Validation Failed

The dns-controller Kubernetes deployment has not updated the Kubernetes cluster's API DNS entry to the correct IP address.  The API DNS IP address is the placeholder address that kops creates: 203.0.113.123.  Please wait about 5-10 minutes for a master to start, dns-controller to launch, and DNS to propagate.  The protokube container and dns-controller deployment logs may contain more diagnostic information.  Etcd and the API DNS entries must be updated for a kops Kubernetes cluster to start.

Validation Failed
W0729 08:05:26.081844    6283 validate_cluster.go:232] (will retry): cluster not yet healthy
INSTANCE GROUPS
NAME			ROLE	MACHINETYPE	MIN	MAX	SUBNETS
master-ap-south-1a	Master	c5.large	1	1	ap-south-1a
nodes-ap-south-1a	Node	t3.medium	4	4	ap-south-1a

NODE STATUS
NAME	ROLE	READY

VALIDATION ERRORS
KIND	NAME		MESSAGE
dns	apiserver	Validation Failed

The dns-controller Kubernetes deployment has not updated the Kubernetes cluster's API DNS entry to the correct IP address.  The API DNS IP address is the placeholder address that kops creates: 203.0.113.123.  Please wait about 5-10 minutes for a master to start, dns-controller to launch, and DNS to propagate.  The protokube container and dns-controller deployment logs may contain more diagnostic information.  Etcd and the API DNS entries must be updated for a kops Kubernetes cluster to start.

Validation Failed
W0729 08:05:36.130963    6283 validate_cluster.go:232] (will retry): cluster not yet healthy
INSTANCE GROUPS
NAME			ROLE	MACHINETYPE	MIN	MAX	SUBNETS
master-ap-south-1a	Master	c5.large	1	1	ap-south-1a
nodes-ap-south-1a	Node	t3.medium	4	4	ap-south-1a

NODE STATUS
NAME	ROLE	READY

VALIDATION ERRORS
KIND	NAME		MESSAGE
dns	apiserver	Validation Failed

The dns-controller Kubernetes deployment has not updated the Kubernetes cluster's API DNS entry to the correct IP address.  The API DNS IP address is the placeholder address that kops creates: 203.0.113.123.  Please wait about 5-10 minutes for a master to start, dns-controller to launch, and DNS to propagate.  The protokube container and dns-controller deployment logs may contain more diagnostic information.  Etcd and the API DNS entries must be updated for a kops Kubernetes cluster to start.

Validation Failed
W0729 08:05:46.170738    6283 validate_cluster.go:232] (will retry): cluster not yet healthy
W0729 08:05:56.218969    6283 validate_cluster.go:184] (will retry): unexpected error during validation: unable to resolve Kubernetes cluster API URL dns: lookup api.e2e-414ec24bfc-41f44.test-cncf-aws.k8s.io on 10.63.240.10:53: no such host
INSTANCE GROUPS
NAME			ROLE	MACHINETYPE	MIN	MAX	SUBNETS
master-ap-south-1a	Master	c5.large	1	1	ap-south-1a
nodes-ap-south-1a	Node	t3.medium	4	4	ap-south-1a

NODE STATUS
NAME	ROLE	READY

VALIDATION ERRORS
KIND	NAME		MESSAGE
dns	apiserver	Validation Failed

The dns-controller Kubernetes deployment has not updated the Kubernetes cluster's API DNS entry to the correct IP address.  The API DNS IP address is the placeholder address that kops creates: 203.0.113.123.  Please wait about 5-10 minutes for a master to start, dns-controller to launch, and DNS to propagate.  The protokube container and dns-controller deployment logs may contain more diagnostic information.  Etcd and the API DNS entries must be updated for a kops Kubernetes cluster to start.

Validation Failed
W0729 08:06:06.271272    6283 validate_cluster.go:232] (will retry): cluster not yet healthy
INSTANCE GROUPS
NAME			ROLE	MACHINETYPE	MIN	MAX	SUBNETS
master-ap-south-1a	Master	c5.large	1	1	ap-south-1a
nodes-ap-south-1a	Node	t3.medium	4	4	ap-south-1a

NODE STATUS
NAME	ROLE	READY

VALIDATION ERRORS
KIND	NAME		MESSAGE
dns	apiserver	Validation Failed

The dns-controller Kubernetes deployment has not updated the Kubernetes cluster's API DNS entry to the correct IP address.  The API DNS IP address is the placeholder address that kops creates: 203.0.113.123.  Please wait about 5-10 minutes for a master to start, dns-controller to launch, and DNS to propagate.  The protokube container and dns-controller deployment logs may contain more diagnostic information.  Etcd and the API DNS entries must be updated for a kops Kubernetes cluster to start.

Validation Failed
W0729 08:06:16.315490    6283 validate_cluster.go:232] (will retry): cluster not yet healthy
INSTANCE GROUPS
NAME			ROLE	MACHINETYPE	MIN	MAX	SUBNETS
master-ap-south-1a	Master	c5.large	1	1	ap-south-1a
nodes-ap-south-1a	Node	t3.medium	4	4	ap-south-1a

NODE STATUS
NAME	ROLE	READY

VALIDATION ERRORS
KIND	NAME		MESSAGE
dns	apiserver	Validation Failed

The dns-controller Kubernetes deployment has not updated the Kubernetes cluster's API DNS entry to the correct IP address.  The API DNS IP address is the placeholder address that kops creates: 203.0.113.123.  Please wait about 5-10 minutes for a master to start, dns-controller to launch, and DNS to propagate.  The protokube container and dns-controller deployment logs may contain more diagnostic information.  Etcd and the API DNS entries must be updated for a kops Kubernetes cluster to start.

Validation Failed
W0729 08:06:26.354409    6283 validate_cluster.go:232] (will retry): cluster not yet healthy
INSTANCE GROUPS
NAME			ROLE	MACHINETYPE	MIN	MAX	SUBNETS
master-ap-south-1a	Master	c5.large	1	1	ap-south-1a
nodes-ap-south-1a	Node	t3.medium	4	4	ap-south-1a

NODE STATUS
NAME	ROLE	READY

VALIDATION ERRORS
KIND	NAME		MESSAGE
dns	apiserver	Validation Failed

The dns-controller Kubernetes deployment has not updated the Kubernetes cluster's API DNS entry to the correct IP address.  The API DNS IP address is the placeholder address that kops creates: 203.0.113.123.  Please wait about 5-10 minutes for a master to start, dns-controller to launch, and DNS to propagate.  The protokube container and dns-controller deployment logs may contain more diagnostic information.  Etcd and the API DNS entries must be updated for a kops Kubernetes cluster to start.

Validation Failed
W0729 08:06:36.387665    6283 validate_cluster.go:232] (will retry): cluster not yet healthy
INSTANCE GROUPS
NAME			ROLE	MACHINETYPE	MIN	MAX	SUBNETS
master-ap-south-1a	Master	c5.large	1	1	ap-south-1a
nodes-ap-south-1a	Node	t3.medium	4	4	ap-south-1a

NODE STATUS
NAME	ROLE	READY

VALIDATION ERRORS
KIND	NAME		MESSAGE
dns	apiserver	Validation Failed

The dns-controller Kubernetes deployment has not updated the Kubernetes cluster's API DNS entry to the correct IP address.  The API DNS IP address is the placeholder address that kops creates: 203.0.113.123.  Please wait about 5-10 minutes for a master to start, dns-controller to launch, and DNS to propagate.  The protokube container and dns-controller deployment logs may contain more diagnostic information.  Etcd and the API DNS entries must be updated for a kops Kubernetes cluster to start.

Validation Failed
W0729 08:06:46.426005    6283 validate_cluster.go:232] (will retry): cluster not yet healthy
INSTANCE GROUPS
NAME			ROLE	MACHINETYPE	MIN	MAX	SUBNETS
master-ap-south-1a	Master	c5.large	1	1	ap-south-1a
nodes-ap-south-1a	Node	t3.medium	4	4	ap-south-1a

NODE STATUS
NAME	ROLE	READY

VALIDATION ERRORS
KIND	NAME		MESSAGE
dns	apiserver	Validation Failed

The dns-controller Kubernetes deployment has not updated the Kubernetes cluster's API DNS entry to the correct IP address.  The API DNS IP address is the placeholder address that kops creates: 203.0.113.123.  Please wait about 5-10 minutes for a master to start, dns-controller to launch, and DNS to propagate.  The protokube container and dns-controller deployment logs may contain more diagnostic information.  Etcd and the API DNS entries must be updated for a kops Kubernetes cluster to start.

Validation Failed
W0729 08:06:56.472305    6283 validate_cluster.go:232] (will retry): cluster not yet healthy
INSTANCE GROUPS
NAME			ROLE	MACHINETYPE	MIN	MAX	SUBNETS
master-ap-south-1a	Master	c5.large	1	1	ap-south-1a
nodes-ap-south-1a	Node	t3.medium	4	4	ap-south-1a

NODE STATUS
NAME	ROLE	READY

VALIDATION ERRORS
KIND	NAME		MESSAGE
dns	apiserver	Validation Failed

The dns-controller Kubernetes deployment has not updated the Kubernetes cluster's API DNS entry to the correct IP address.  The API DNS IP address is the placeholder address that kops creates: 203.0.113.123.  Please wait about 5-10 minutes for a master to start, dns-controller to launch, and DNS to propagate.  The protokube container and dns-controller deployment logs may contain more diagnostic information.  Etcd and the API DNS entries must be updated for a kops Kubernetes cluster to start.

Validation Failed
W0729 08:07:06.505428    6283 validate_cluster.go:232] (will retry): cluster not yet healthy
INSTANCE GROUPS
NAME			ROLE	MACHINETYPE	MIN	MAX	SUBNETS
master-ap-south-1a	Master	c5.large	1	1	ap-south-1a
nodes-ap-south-1a	Node	t3.medium	4	4	ap-south-1a

NODE STATUS
NAME	ROLE	READY

VALIDATION ERRORS
KIND	NAME		MESSAGE
dns	apiserver	Validation Failed

The dns-controller Kubernetes deployment has not updated the Kubernetes cluster's API DNS entry to the correct IP address.  The API DNS IP address is the placeholder address that kops creates: 203.0.113.123.  Please wait about 5-10 minutes for a master to start, dns-controller to launch, and DNS to propagate.  The protokube container and dns-controller deployment logs may contain more diagnostic information.  Etcd and the API DNS entries must be updated for a kops Kubernetes cluster to start.

Validation Failed
W0729 08:07:16.556711    6283 validate_cluster.go:232] (will retry): cluster not yet healthy
INSTANCE GROUPS
NAME			ROLE	MACHINETYPE	MIN	MAX	SUBNETS
master-ap-south-1a	Master	c5.large	1	1	ap-south-1a
nodes-ap-south-1a	Node	t3.medium	4	4	ap-south-1a

NODE STATUS
NAME	ROLE	READY

VALIDATION ERRORS
KIND	NAME		MESSAGE
dns	apiserver	Validation Failed

The dns-controller Kubernetes deployment has not updated the Kubernetes cluster's API DNS entry to the correct IP address.  The API DNS IP address is the placeholder address that kops creates: 203.0.113.123.  Please wait about 5-10 minutes for a master to start, dns-controller to launch, and DNS to propagate.  The protokube container and dns-controller deployment logs may contain more diagnostic information.  Etcd and the API DNS entries must be updated for a kops Kubernetes cluster to start.

Validation Failed
W0729 08:07:26.599852    6283 validate_cluster.go:232] (will retry): cluster not yet healthy
INSTANCE GROUPS
NAME			ROLE	MACHINETYPE	MIN	MAX	SUBNETS
master-ap-south-1a	Master	c5.large	1	1	ap-south-1a
nodes-ap-south-1a	Node	t3.medium	4	4	ap-south-1a

NODE STATUS
NAME	ROLE	READY

VALIDATION ERRORS
KIND	NAME		MESSAGE
dns	apiserver	Validation Failed

The dns-controller Kubernetes deployment has not updated the Kubernetes cluster's API DNS entry to the correct IP address.  The API DNS IP address is the placeholder address that kops creates: 203.0.113.123.  Please wait about 5-10 minutes for a master to start, dns-controller to launch, and DNS to propagate.  The protokube container and dns-controller deployment logs may contain more diagnostic information.  Etcd and the API DNS entries must be updated for a kops Kubernetes cluster to start.

Validation Failed
W0729 08:07:36.647377    6283 validate_cluster.go:232] (will retry): cluster not yet healthy
INSTANCE GROUPS
NAME			ROLE	MACHINETYPE	MIN	MAX	SUBNETS
master-ap-south-1a	Master	c5.large	1	1	ap-south-1a
nodes-ap-south-1a	Node	t3.medium	4	4	ap-south-1a

NODE STATUS
NAME	ROLE	READY

VALIDATION ERRORS
KIND	NAME		MESSAGE
dns	apiserver	Validation Failed

The dns-controller Kubernetes deployment has not updated the Kubernetes cluster's API DNS entry to the correct IP address.  The API DNS IP address is the placeholder address that kops creates: 203.0.113.123.  Please wait about 5-10 minutes for a master to start, dns-controller to launch, and DNS to propagate.  The protokube container and dns-controller deployment logs may contain more diagnostic information.  Etcd and the API DNS entries must be updated for a kops Kubernetes cluster to start.

Validation Failed
W0729 08:07:46.677593    6283 validate_cluster.go:232] (will retry): cluster not yet healthy
INSTANCE GROUPS
NAME			ROLE	MACHINETYPE	MIN	MAX	SUBNETS
master-ap-south-1a	Master	c5.large	1	1	ap-south-1a
nodes-ap-south-1a	Node	t3.medium	4	4	ap-south-1a

NODE STATUS
NAME	ROLE	READY

VALIDATION ERRORS
KIND	NAME		MESSAGE
dns	apiserver	Validation Failed

The dns-controller Kubernetes deployment has not updated the Kubernetes cluster's API DNS entry to the correct IP address.  The API DNS IP address is the placeholder address that kops creates: 203.0.113.123.  Please wait about 5-10 minutes for a master to start, dns-controller to launch, and DNS to propagate.  The protokube container and dns-controller deployment logs may contain more diagnostic information.  Etcd and the API DNS entries must be updated for a kops Kubernetes cluster to start.

Validation Failed
W0729 08:07:56.725355    6283 validate_cluster.go:232] (will retry): cluster not yet healthy
INSTANCE GROUPS
NAME			ROLE	MACHINETYPE	MIN	MAX	SUBNETS
master-ap-south-1a	Master	c5.large	1	1	ap-south-1a
nodes-ap-south-1a	Node	t3.medium	4	4	ap-south-1a

NODE STATUS
NAME	ROLE	READY

VALIDATION ERRORS
KIND	NAME		MESSAGE
dns	apiserver	Validation Failed

The dns-controller Kubernetes deployment has not updated the Kubernetes cluster's API DNS entry to the correct IP address.  The API DNS IP address is the placeholder address that kops creates: 203.0.113.123.  Please wait about 5-10 minutes for a master to start, dns-controller to launch, and DNS to propagate.  The protokube container and dns-controller deployment logs may contain more diagnostic information.  Etcd and the API DNS entries must be updated for a kops Kubernetes cluster to start.

Validation Failed
W0729 08:08:06.756762    6283 validate_cluster.go:232] (will retry): cluster not yet healthy
INSTANCE GROUPS
NAME			ROLE	MACHINETYPE	MIN	MAX	SUBNETS
master-ap-south-1a	Master	c5.large	1	1	ap-south-1a
nodes-ap-south-1a	Node	t3.medium	4	4	ap-south-1a

... skipping 29 lines ...
Pod	kube-system/metrics-server-7c9d469d74-2j82x				system-cluster-critical pod "metrics-server-7c9d469d74-2j82x" is pending
Pod	kube-system/metrics-server-7c9d469d74-dw8wg				system-cluster-critical pod "metrics-server-7c9d469d74-dw8wg" is pending
Pod	kube-system/node-local-dns-65j7g					system-node-critical pod "node-local-dns-65j7g" is pending
Pod	kube-system/node-local-dns-7zklr					system-node-critical pod "node-local-dns-7zklr" is pending
Pod	kube-system/node-local-dns-k5jtz					system-node-critical pod "node-local-dns-k5jtz" is pending

Validation Failed
W0729 08:08:22.248582    6283 validate_cluster.go:232] (will retry): cluster not yet healthy
INSTANCE GROUPS
NAME			ROLE	MACHINETYPE	MIN	MAX	SUBNETS
master-ap-south-1a	Master	c5.large	1	1	ap-south-1a
nodes-ap-south-1a	Node	t3.medium	4	4	ap-south-1a

... skipping 24 lines ...
Pod	kube-system/cluster-autoscaler-84864c677f-llscr			system-cluster-critical pod "cluster-autoscaler-84864c677f-llscr" is pending
Pod	kube-system/coredns-78cd66cbc9-s724g				system-cluster-critical pod "coredns-78cd66cbc9-s724g" is pending
Pod	kube-system/coredns-autoscaler-6d96c59bbf-fgbth			system-cluster-critical pod "coredns-autoscaler-6d96c59bbf-fgbth" is pending
Pod	kube-system/metrics-server-7c9d469d74-2j82x			system-cluster-critical pod "metrics-server-7c9d469d74-2j82x" is pending
Pod	kube-system/metrics-server-7c9d469d74-dw8wg			system-cluster-critical pod "metrics-server-7c9d469d74-dw8wg" is pending

Validation Failed
W0729 08:08:36.193591    6283 validate_cluster.go:232] (will retry): cluster not yet healthy
INSTANCE GROUPS
NAME			ROLE	MACHINETYPE	MIN	MAX	SUBNETS
master-ap-south-1a	Master	c5.large	1	1	ap-south-1a
nodes-ap-south-1a	Node	t3.medium	4	4	ap-south-1a

... skipping 18 lines ...
Pod	kube-system/cilium-tmvvd					system-node-critical pod "cilium-tmvvd" is not ready (cilium-agent)
Pod	kube-system/coredns-78cd66cbc9-s724g				system-cluster-critical pod "coredns-78cd66cbc9-s724g" is pending
Pod	kube-system/coredns-78cd66cbc9-stfds				system-cluster-critical pod "coredns-78cd66cbc9-stfds" is pending
Pod	kube-system/metrics-server-7c9d469d74-2j82x			system-cluster-critical pod "metrics-server-7c9d469d74-2j82x" is not ready (metrics-server)
Pod	kube-system/metrics-server-7c9d469d74-dw8wg			system-cluster-critical pod "metrics-server-7c9d469d74-dw8wg" is pending

Validation Failed
W0729 08:08:50.090629    6283 validate_cluster.go:232] (will retry): cluster not yet healthy
INSTANCE GROUPS
NAME			ROLE	MACHINETYPE	MIN	MAX	SUBNETS
master-ap-south-1a	Master	c5.large	1	1	ap-south-1a
nodes-ap-south-1a	Node	t3.medium	4	4	ap-south-1a

... skipping 11 lines ...
Pod	kube-system/cert-manager-7bc755847-t4hl9			system-cluster-critical pod "cert-manager-7bc755847-t4hl9" is pending
Pod	kube-system/cert-manager-webhook-9b4bf487-t78j8			system-cluster-critical pod "cert-manager-webhook-9b4bf487-t78j8" is pending
Pod	kube-system/cilium-tmvvd					system-node-critical pod "cilium-tmvvd" is not ready (cilium-agent)
Pod	kube-system/metrics-server-7c9d469d74-2j82x			system-cluster-critical pod "metrics-server-7c9d469d74-2j82x" is not ready (metrics-server)
Pod	kube-system/metrics-server-7c9d469d74-dw8wg			system-cluster-critical pod "metrics-server-7c9d469d74-dw8wg" is not ready (metrics-server)

Validation Failed
W0729 08:09:03.968660    6283 validate_cluster.go:232] (will retry): cluster not yet healthy
INSTANCE GROUPS
NAME			ROLE	MACHINETYPE	MIN	MAX	SUBNETS
master-ap-south-1a	Master	c5.large	1	1	ap-south-1a
nodes-ap-south-1a	Node	t3.medium	4	4	ap-south-1a

... skipping 8 lines ...
VALIDATION ERRORS
KIND	NAME								MESSAGE
Pod	kube-system/aws-load-balancer-controller-8569cf8cf9-42lfm	system-cluster-critical pod "aws-load-balancer-controller-8569cf8cf9-42lfm" is pending
Pod	kube-system/cilium-tmvvd					system-node-critical pod "cilium-tmvvd" is not ready (cilium-agent)
Pod	kube-system/metrics-server-7c9d469d74-2j82x			system-cluster-critical pod "metrics-server-7c9d469d74-2j82x" is not ready (metrics-server)

Validation Failed
W0729 08:09:17.810630    6283 validate_cluster.go:232] (will retry): cluster not yet healthy
INSTANCE GROUPS
NAME			ROLE	MACHINETYPE	MIN	MAX	SUBNETS
master-ap-south-1a	Master	c5.large	1	1	ap-south-1a
nodes-ap-south-1a	Node	t3.medium	4	4	ap-south-1a

... skipping 6 lines ...
ip-172-20-0-242.ap-south-1.compute.internal	node	True

VALIDATION ERRORS
KIND	NAME								MESSAGE
Pod	kube-system/aws-load-balancer-controller-8569cf8cf9-42lfm	system-cluster-critical pod "aws-load-balancer-controller-8569cf8cf9-42lfm" is pending

Validation Failed
W0729 08:09:31.558521    6283 validate_cluster.go:232] (will retry): cluster not yet healthy
INSTANCE GROUPS
NAME			ROLE	MACHINETYPE	MIN	MAX	SUBNETS
master-ap-south-1a	Master	c5.large	1	1	ap-south-1a
nodes-ap-south-1a	Node	t3.medium	4	4	ap-south-1a

... skipping 6 lines ...
ip-172-20-0-242.ap-south-1.compute.internal	node	True

VALIDATION ERRORS
KIND	NAME								MESSAGE
Pod	kube-system/aws-load-balancer-controller-8569cf8cf9-42lfm	system-cluster-critical pod "aws-load-balancer-controller-8569cf8cf9-42lfm" is pending

Validation Failed
W0729 08:09:45.331150    6283 validate_cluster.go:232] (will retry): cluster not yet healthy
INSTANCE GROUPS
NAME			ROLE	MACHINETYPE	MIN	MAX	SUBNETS
master-ap-south-1a	Master	c5.large	1	1	ap-south-1a
nodes-ap-south-1a	Node	t3.medium	4	4	ap-south-1a

... skipping 6 lines ...
ip-172-20-0-242.ap-south-1.compute.internal	node	True

VALIDATION ERRORS
KIND	NAME								MESSAGE
Pod	kube-system/aws-load-balancer-controller-8569cf8cf9-42lfm	system-cluster-critical pod "aws-load-balancer-controller-8569cf8cf9-42lfm" is pending

Validation Failed
W0729 08:09:59.288768    6283 validate_cluster.go:232] (will retry): cluster not yet healthy
INSTANCE GROUPS
NAME			ROLE	MACHINETYPE	MIN	MAX	SUBNETS
master-ap-south-1a	Master	c5.large	1	1	ap-south-1a
nodes-ap-south-1a	Node	t3.medium	4	4	ap-south-1a

... skipping 6 lines ...
ip-172-20-0-242.ap-south-1.compute.internal	node	True

VALIDATION ERRORS
KIND	NAME								MESSAGE
Pod	kube-system/aws-load-balancer-controller-8569cf8cf9-42lfm	system-cluster-critical pod "aws-load-balancer-controller-8569cf8cf9-42lfm" is pending

Validation Failed
W0729 08:10:13.033775    6283 validate_cluster.go:232] (will retry): cluster not yet healthy
INSTANCE GROUPS
NAME			ROLE	MACHINETYPE	MIN	MAX	SUBNETS
master-ap-south-1a	Master	c5.large	1	1	ap-south-1a
nodes-ap-south-1a	Node	t3.medium	4	4	ap-south-1a

... skipping 1009 lines ...
I0729 08:15:26.328371    6384 channel_version.go:140] manifest Match for "node-termination-handler.aws": Channel=s3://k8s-kops-prow/e2e-414ec24bfc-41f44.test-cncf-aws.k8s.io/addons/bootstrap-channel.yaml Id=k8s-1.11 ManifestHash=ecd9196336f783cd872f6ff736bb082aadcd6cf9eee27de17b8a4324fee1dd65 SystemGeneration=1
NAME			CURRENT	UPDATE									PKI
networking.cilium.io	-	0d7e665f356f97c0137cda33453dbd5d40d5fbca30b62df8ef30e2db894a36a1	no
I0729 08:15:26.825862    6384 addon.go:188] Applying update from "s3://k8s-kops-prow/e2e-414ec24bfc-41f44.test-cncf-aws.k8s.io/addons/networking.cilium.io/k8s-1.16-v1.11.yaml"
I0729 08:15:26.825917    6384 s3fs.go:329] Reading file "s3://k8s-kops-prow/e2e-414ec24bfc-41f44.test-cncf-aws.k8s.io/addons/networking.cilium.io/k8s-1.16-v1.11.yaml"
I0729 08:15:26.937869    6384 apply.go:83] Running command: kubectl apply -f /tmp/channel950785448/manifest.yaml --server-side --force-conflicts --field-manager=kops
I0729 08:15:33.540332    6384 apply.go:86] error running kubectl apply -f /tmp/channel950785448/manifest.yaml --server-side --force-conflicts --field-manager=kops
I0729 08:15:33.540397    6384 apply.go:87] serviceaccount/cilium serverside-applied
serviceaccount/cilium-operator serverside-applied
serviceaccount/hubble-relay serverside-applied
configmap/cilium-config serverside-applied
configmap/hubble-relay-config serverside-applied
clusterrole.rbac.authorization.k8s.io/cilium serverside-applied
... skipping 5 lines ...
service/hubble-relay serverside-applied
daemonset.apps/cilium serverside-applied
deployment.apps/cilium-operator serverside-applied
certificate.cert-manager.io/hubble-server-certs serverside-applied
certificate.cert-manager.io/hubble-relay-client-certs serverside-applied
poddisruptionbudget.policy/cilium-operator serverside-applied
Error from server: failed to create typed patch object: .spec.template.spec.containers[name="hubble-relay"].ports: element 0: associative list with keys has an element that omits key field "protocol" (and doesn't have default value)
E0729 08:15:33.540418    6384 apply.go:52] failed to apply the manifest: error running kubectl: exit status 1
I0729 08:15:33.540507    6384 apply.go:83] Running command: kubectl replace -f /tmp/channel950785448/manifest.yaml --field-manager=kops
I0729 08:15:43.531262    6384 apply.go:86] error running kubectl replace -f /tmp/channel950785448/manifest.yaml --field-manager=kops
I0729 08:15:43.531304    6384 apply.go:87] serviceaccount/cilium replaced
serviceaccount/cilium-operator replaced
serviceaccount/hubble-relay replaced
configmap/cilium-config replaced
configmap/hubble-relay-config replaced
clusterrole.rbac.authorization.k8s.io/cilium replaced
... skipping 4 lines ...
clusterrolebinding.rbac.authorization.k8s.io/hubble-relay replaced
daemonset.apps/cilium replaced
deployment.apps/cilium-operator replaced
certificate.cert-manager.io/hubble-server-certs replaced
certificate.cert-manager.io/hubble-relay-client-certs replaced
poddisruptionbudget.policy/cilium-operator replaced
Error from server (Invalid): error when replacing "/tmp/channel950785448/manifest.yaml": Service "hubble-relay" is invalid: spec.clusterIP: Invalid value: "": field is immutable
Error from server (NotFound): error when replacing "/tmp/channel950785448/manifest.yaml": deployments.apps "hubble-relay" not found
E0729 08:15:43.531322    6384 apply.go:61] failed to replace manifest: error running kubectl: exit status 1
I0729 08:15:43.531400    6384 apply.go:83] Running command: kubectl apply -f /tmp/channel950785448/manifest.yaml --server-side --force-conflicts --field-manager=kops
I0729 08:15:49.070708    6384 apply.go:86] error running kubectl apply -f /tmp/channel950785448/manifest.yaml --server-side --force-conflicts --field-manager=kops
I0729 08:15:49.070763    6384 apply.go:87] serviceaccount/cilium serverside-applied
serviceaccount/cilium-operator serverside-applied
serviceaccount/hubble-relay serverside-applied
configmap/cilium-config serverside-applied
configmap/hubble-relay-config serverside-applied
clusterrole.rbac.authorization.k8s.io/cilium serverside-applied
... skipping 5 lines ...
service/hubble-relay serverside-applied
daemonset.apps/cilium serverside-applied
deployment.apps/cilium-operator serverside-applied
certificate.cert-manager.io/hubble-server-certs serverside-applied
certificate.cert-manager.io/hubble-relay-client-certs serverside-applied
poddisruptionbudget.policy/cilium-operator serverside-applied
Error from server: failed to create typed patch object: .spec.template.spec.containers[name="hubble-relay"].ports: element 0: associative list with keys has an element that omits key field "protocol" (and doesn't have default value)

updating "networking.cilium.io": error applying update from "s3://k8s-kops-prow/e2e-414ec24bfc-41f44.test-cncf-aws.k8s.io/addons/networking.cilium.io/k8s-1.16-v1.11.yaml": failed to apply the manifest: error running kubectl: exit status 1
+ kops-finish
+ kubetest2 kops -v=2 --cloud-provider=aws --cluster-name=e2e-414ec24bfc-41f44.test-cncf-aws.k8s.io --kops-root=/home/prow/go/src/k8s.io/kops --admin-access= --env=KOPS_FEATURE_FLAGS=SpecOverrideFlag --kops-binary-path=/tmp/kops.EvnerqtlZ --down
I0729 08:15:49.099990    6432 featureflag.go:162] FeatureFlag "SpecOverrideFlag"=true
I0729 08:15:49.101144    6432 app.go:62] The files in RunDir shall not be part of Artifacts
I0729 08:15:49.101178    6432 app.go:63] pass rundir-in-artifacts flag True for RunDir to be part of Artifacts
I0729 08:15:49.101206    6432 app.go:65] RunDir for this run: "/home/prow/go/src/k8s.io/kops/_rundir/871c3a8b-0f14-11ed-b6e4-d28c295d3c4e"
... skipping 304 lines ...