-
Notifications
You must be signed in to change notification settings - Fork 4.7k
New issue
Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.
By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.
Already on GitHub? Sign in to your account
Use a dedicated etcd cluster for Cilium CNI #6496
Comments
the bundled addon is also still using cilium 1.0-stable instead of 1.4 EDIT: seems EDIT2: the issues we had with currently bundled Cilium templates are the ds missing features and rbac missing permissions. the suggestion to use cilium-etcd-operator by default should come with a warning that this is meant for "small" clusters, doesn't consider IOPS requirements for etcd and etcd failures may result in seconds of network instability as cilium recovers the state into etcd |
Yep would be great to have a working solution for the upcoming 1.12 that works with more than just "small" clusters. cc @nebril |
The way forward would be to add cilium-etcd-operator to the ciliumkops add-on. The problem is that during cluster creation, only master nodes are available. And cilium-etcd-operator and the pods that they spawn currently don't have the tolerations necessary for them to be scheduled on master nodes. So that has to be added before anything can be done on the kops side. For new clusters using etcd-manager, it won't be possible to use cilium at all because cilium running on normal nodes won't be able to talk to kops etcd anymore. For existing clusters, we have managed to migrate to using cilium-etcd-operator just fine as then they can just be scheduled on normal nodes and we change the cilium k/v store before switching to etcd-manager. |
Ah that is good to know @olemarkus and you are not experiencing any of the performance/scalibility/reliability issue sthat @so0k is referring to? And the tolerations issue is cilium/cilium-etcd-operator#42 I presume? |
Just had a chat with @tgraf and with cilium 1.6 there will also be CRD support: cilium/cilium#7573 which would make it possible to run without etcd when the cluster is below a certain size. |
Performance-wise it should be more than enough. This is also what cilium use for scalability testing, I think. Reliability ... we have seen etcd cluster has been recreated during rolling updates.This has been improving quite a bit though.
Yes. |
I tried installing cilium with etcd operator manually using |
Given that virtually no resources match the ones that kops create anymore, I think it makes sense to set That will also ensure kops does not override these resources unexpectedly in the future. |
Since Kops is locking down etcd with version 1.12, Cilium should start using a dedicated etcd cluster. One option is to use the cilium-etcd-operator that ships with Cilium 1.4. Example resources for provisioning this can be found here: https://github.com/cilium/cilium/tree/master/examples/kubernetes/1.11
The text was updated successfully, but these errors were encountered: