-
Notifications
You must be signed in to change notification settings - Fork 1.9k
New issue
Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.
By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.
Already on GitHub? Sign in to your account
[argo-cd] Enabling ha with autoscaling results in redis-ha-haproxy crashing with OOMKilled #1958
Comments
Could it be haproxy specific? What image is being used for redis-ha-haproxy? Maybe this? |
Hmm your mentioned chart version include haproxy 2.6.4, the issue you mentioned targets 2.6.3. Can you try to add some limits on the haproxy pods? Eg. redis-ha:
haproxy:
resources:
limits:
cpu: 1
memory: 512Mi |
Am trying out different versions of haproxy using:
Reference: |
Looks like kubernetes relies on this to be fixed at the container service level, in my case this is containerd fixed like this:
|
I do not understand how «out of memory» is related to «limit of number of open files» ? |
Adding this comment reference as elaboration source: kubernetes/kubernetes#3595 (comment) |
Creating a systemd drop-in via ansible relaxed the situation for me. Setting the process defaults of soft 1024 and hard 524288 for the containerd service. - name: Set RHEL9 ulimit
hosts: all
tasks:
- name: Create dropin directory if not exists
ansible.builtin.file:
path: /etc/systemd/system/containerd.service.d
state: directory
when:
- ansible_os_family == "RedHat"
- ansible_distribution_major_version == "9"
- name: Add ulimits dropin
ansible.builtin.copy:
dest: /etc/systemd/system/containerd.service.d/ulimits.conf
content: |
[Service]
LimitNOFILE=
LimitNOFILE=1024:524288
when:
- ansible_os_family == "RedHat"
- ansible_distribution_major_version == "9"
notify:
- Restart containerd
handlers:
- name: Restart containerd
ansible.builtin.systemd_service:
daemon_reload: true
name: containerd
state: restarted
enabled: true |
Describe the bug
Set values.yaml as described to enable ha with autoscaling here:
https://github.com/argoproj/argo-helm/tree/main/charts/argo-cd
The redis-ha-haproxy pods are crashing with OOMKilled.
Related helm chart
argo-cd
Helm chart version
5.28.2
To Reproduce
Expected behavior
argocd stood up with ha configuration
Screenshots
Additional context
kubernetes v1.25.8
os redhat 9
k8s installed via kubeadm
I boosted the memory by 4 gb at a time up to 48 on each of 3 worker nodes, this is a newly setup cluster with this argocd deployment pretty much the only thing running. If I ssh into each of the worker nodes it shows haproxy using up all the cpu and memory.
The text was updated successfully, but these errors were encountered: