-
Notifications
You must be signed in to change notification settings - Fork 2k
New issue
Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.
By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.
Already on GitHub? Sign in to your account
Nomad in small clusters #2176
Comments
Hey @kak-tus, Nomad will not restart jobs just because of a leader election. What is the ping time between the servers and can you share the logs of the servers/clients after said leader transistion/restarting jobs. |
@dadgar Hm, you are right. As I remember 0.5.0 was more stable. May be something was changed in 0.5.1 or 0.5.2. But may be network stability was changed. Normal ping between servers - 1.5-2.5 ms. Aggregated log of a hole cluster (c1,c2,c3 in log - nodes). |
I reverted back to 0.5.0 and will be seen cluster behavior. |
@kak-tus I am going to close this is Nomad does not behave in the way described in the issue. Further the logs do show large latency between the servers. It may have just been a transient network issue |
I'm going to lock this issue because it has been closed for 120 days ⏳. This helps our maintainers find and focus on the active issues. |
I have two nomad clusters: one production cluster at work with enough nodes, and second - "just for fun" cluster, for my private services.
First cluster with many nodes work without any trouble.
Small cluster has only 3 nodes. Each node has server and client and placed in different datacenter (so I have 1 region and 3 DC in nomad terminology).
And this small cluster is not very stable to network lags.
Masters reelecting frequently (every 1-2 days) because of temporary network lags, but it is not very bad.
Bad in this case: after masters reelection nomad begin to restart every job in cluster.
But I like nomad and I want to use it in my small cluster.
How I see fixing this problem:
The text was updated successfully, but these errors were encountered: