-
Notifications
You must be signed in to change notification settings - Fork 2.9k
New issue
Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.
By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.
Already on GitHub? Sign in to your account
Flannel POD OOMKilled #963
Comments
We have updated resources limits and there are no more restart. So the question is :
|
I'm having the same issue. |
same here |
The same issue for me. Fixed by setting 256Mi memory limit in flannel daemonSet properties |
I'm observing a very similar issue.
I didn't observe these issues with flannel
@tomdee do you think this issue may be caused by the memory limit being too low? (as advised by @Tapolsky above) Here's the status of this OOMKilled pod:
And status on the node:
|
@ttarczynski I guess recent releases of flannel consume more memory, flannel-0.10 changed a lot on vxlan backend. I also got this OOMKilled issue and I increased its memory limit from 50Mi to 100Mi. You are free to increase the memory limit, the original purpose in #855 is to explicitly set resource requests and limits to get QoS class "Guaranteed". But better not increase too much, it's fine to oom-kill flannel if it's abnormal, network connectivity won't be affected if flanneld restarts quickly. @tomdee maybe it's time to increase the default memory limit? or better to profile why it consumes so much memory? My cluster has less than ten k8s worker nodes. |
@Dieken In my case I've already increased the limit from 50Mi to 100Mi and still got it OOMKilled. My cluster is also only 10 nodes and I always see the |
@ttarczynski I double checked my clusters, luckily flannel pods didn't restart due to OOMKilled, rarely they restarted with error code 255, possibly due to some kube-apiserver error. Because we both have small scale cluster, I doubt flanneld-0.10 introduced some defect. BTW, you may try old version, I use k8s-1.8.13 + flannel-0.9.0 and k8s-1.9.8 + flannel-0.9.1, both with 100Mi memory limit for flannel, and actual memory usage below 30Mi when I just checked. |
The same problem. Rollback to 0.9.1 fix it. |
I think the problems I've seen were related to the default kernel version in CentOS 7 (
And probably this was causing flannel to be OOMKilled in my cluster. |
I have upgraded the kernel to But still i am facing Any know fixes for this ? Does increasing the memory limit actually solves this issue ? Interestingly enough i am facing this issue only in server class machines! More details:Kubernetes detailsSystem logs |
Maybe you can modify the memory limit, and test it. kubectl patch ds -n=kube-system kube-flannel-ds-amd64 -p '{"spec": {"template":{"spec":{"containers": [{"name":"kube-flannel", "resources": {"limits": {"cpu": "250m","memory": "550Mi"},"requests": {"cpu": "100m","memory": "100Mi"}}}]}}}}' |
@lubinsz Yes increased the memory limit. Now it works fine Thanks |
same problem
go to that node
have change limit, Need further observation |
This issue has been automatically marked as stale because it has not had recent activity. It will be closed if no further activity occurs. Thank you for your contributions. |
The same issue for me. Fixed by setting 256Mi memory limit in flannel daemonSet properties |
Flannels pod have 600 restart on node only (not on master) due to OOMKilled.
Expected Behavior
No Restart
Possible Solution
Increase memory limit
Steps to Reproduce (for bugs)
install flannel using prodived yaml (Documentation/kube-flannel.yml)
Context
1 Master / 2 Node
Your Environment
What's the recommended memory limit for Flannel ?
The text was updated successfully, but these errors were encountered: