-
Notifications
You must be signed in to change notification settings - Fork 3k
New issue
Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.
By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.
Already on GitHub? Sign in to your account
InfluxDB Consuming A Lot of Memory #9527
Comments
I believe the add ons needs a version bump for influx db. It currently uses 1.1.1 and we should have 1.2.0+ as that contains the fixes here: influxdata/influxdb#7832 |
@aemneina @Joeskyyy for influxdb we use images from the official kubernetes repo. They do build it and only use certain packages of the official influxdb repo (https://github.com/influxdata/influxdb). The version of those packages is v.0.12.2 (https://github.com/kubernetes/heapster/blob/master/Godeps/Godeps.json#L241). That eliminates a possibility of a memory leak that was introduced later, in 1.1.0 and 1.1.1. So the bug can be as well in k8s heapster code base, and we need to file for https://github.com/kubernetes/kubernetes/issues. Not sure if there are any profiling tools enabled on the image so we could take the memory footprint to make it easier to debug for them. |
@alena1108 it looks like the addon starter is hard coded to use an older images for influx (1.1.1), the official k8s repo seems to be using v1.3.3 now. We should update the addon starter here. Heapster and grafana also have version bumps. |
@moelsayed Can you confirm what versions will be used in k8s 1.8? is the later version for influx compatible for the currently released k8s 1.7? |
@deniseschannon the version we updated to in 1.8 is heapster-influxdb-amd64:v1.3.3. I am not 100% sure if its compatible with 1.7. We probably need to test that to confirm. |
@deniseschannon I tested and the updated stacks we are using with k8s 1.8 work fine on the released k8s 1.7. |
Updated versions of the addons are already merged, tested and working. The new versions should fix this issue:
|
Rancher versions:
rancher/server: v1.6.0
rancher/agent :v0.5.4
Infrastructure Stack versions:
healthcheck: v0.3.1
ipsec: v0.1.0
network-services: v0.2.0
scheduler: v0.5.1
kubernetes (if applicable): v1.5.4-rancher1-4.1
Docker version: (
docker version
,docker info
preferred)Operating system and kernel: (
cat /etc/os-release
,uname -r
preferred)Type/provider of hosts: (VirtualBox/Bare-metal/AWS/GCE/DO)
AWS Nodes
Setup details: (single node rancher vs. HA rancher, internal DB vs. external DB)
HA rancher with external DB
Environment Template: (Cattle/Kubernetes/Swarm/Mesos)
Kubernetes
Steps to Reproduce:
Let the cluster sit for a bit, launch some deployments and run some workloads.
Results:
InfluxDB will consume a bunch of memory
This particular cluster only has 121 containers running (that's including all the kubernetes/rancher goods), 4 AWS hosts to host those.
Talked with Ahmad in slack, they asked me to file an issue for it for discussion.
The text was updated successfully, but these errors were encountered: