-
Notifications
You must be signed in to change notification settings - Fork 40k
New issue
Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.
By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.
Already on GitHub? Sign in to your account
kube-controller-manager memory leak #112319
Comments
/sig api-machinery |
pmap command report memory map of |
/cc @jpbetz |
that is an old version, you need to move to one of the supported versions. golang pprof should give you a better understanding of the problem |
@aojea |
@aojea When the memory usage of use the |
Could you provide CPU profile? It may be usefully, @xigang |
I don't think this issue is related to issue #102565 issue #102718. issue #102565 #102718 reflects the use of a lot of memory when doing ListObject. And from the kube-controller-manager you provided, the memory-related pprof consumes a lot of memory in the Unmarshal of a separate Object (*corev1.Node). FYI: My guess here is because a large number of Nodes in the kube cluster briefly jittered. |
emm,,, did this use of a lot of memory go on for a long time? |
This problem lasted for a few days and currently the average cluster |
According to the information provided so far (heap-pprof inuse_space): We can draw the following conclusions:
I noticed that the version of golang used by kube-controller-manager on your side is: Suggest:
|
@sxllwx Thank you very much, the memory usage of |
The Kubernetes project currently lacks enough contributors to adequately respond to all issues and PRs. This bot triages issues and PRs according to the following rules:
You can:
Please send feedback to sig-contributor-experience at kubernetes/community. /lifecycle stale |
/remove-lifecycle stale |
/close |
@xigang: Closing this issue. In response to this:
Instructions for interacting with me using PR comments are available here. If you have questions or suggestions related to my behavior, please file an issue against the kubernetes/test-infra repository. |
Hi @xigang , I am interest in this prombel. What have you done before it become normal? |
@lowang-bh The daemonset controller processed Node events slowly, triggering the OOM. We made some optimization. |
/reopen |
@xigang: Reopened this issue. In response to this:
Instructions for interacting with me using PR comments are available here. If you have questions or suggestions related to my behavior, please file an issue against the kubernetes/test-infra repository. |
Have you currently adopted this advice I gave? |
These policies are not currently being adjusted. |
/close |
@xigang: Closing this issue. In response to this:
Instructions for interacting with me using PR comments are available here. If you have questions or suggestions related to my behavior, please file an issue against the kubernetes/test-infra repository. |
What happened?
Our cluster has 5600 nodes and
kube-controller-manager
memory usage 197.5GiB, After restartingkube-controller-manager
, the memory usage is only 15GiB, It looks like the controller manager memory is leaking.The same problem issue: #102718 and #102565
What did you expect to happen?
The
kube-controller-manager
memory usage should be kept at a smooth water level.How can we reproduce it (as minimally and precisely as possible)?
NONE
Anything else we need to know?
NONE
Kubernetes version
Cloud provider
OS version
Install tools
NONE
Container runtime (CRI) and version (if applicable)
Related plugins (CNI, CSI, ...) and versions (if applicable)
The text was updated successfully, but these errors were encountered: