-
Notifications
You must be signed in to change notification settings - Fork 300
zombies #276
Comments
I checked all the MLBs in my soak cluster and I'm not seeing this: root@ip-10-0-6-34:/marathon-lb# ps waux
USER PID %CPU %MEM VSZ RSS TTY STAT START TIME COMMAND
root 1 0.0 0.0 20256 3060 ? Ss 15:46 0:00 /bin/bash /marathon-lb/run sse -m http://master.mesos:8080 --health-check --haproxy
root 8 0.0 0.0 4088 712 ? S 15:46 0:00 /usr/bin/runsv /marathon-lb/service/haproxy
root 9 0.1 0.1 142676 23500 ? Sl 15:46 0:00 python3 /marathon-lb/marathon_lb.py --syslog-socket /dev/null --haproxy-config /mar
root 10 0.0 0.0 20264 3068 ? S 15:46 0:00 /bin/bash ./run
root 490 0.1 0.0 40556 11556 ? Ss 15:46 0:00 haproxy -p /tmp/haproxy.pid -f /marathon-lb/haproxy.cfg -D -sf 252
root 1255 0.0 0.0 20332 3360 ? Ss+ 15:52 0:00 /bin/bash
root 1325 0.3 0.0 20332 3356 ? Ss 15:52 0:00 /bin/bash
root 1343 0.0 0.0 4224 716 ? S 15:52 0:00 sleep 0.5
root 1344 0.0 0.0 34492 2848 ? R+ 15:52 0:00 ps waux
root@ip-10-0-6-34:/marathon-lb# ps waux
USER PID %CPU %MEM VSZ RSS TTY STAT START TIME COMMAND
root 1 0.0 0.0 20256 3060 ? Ss 15:46 0:00 /bin/bash /marathon-lb/run sse -m http://master.mesos:8080 --health-check --haproxy
root 8 0.0 0.0 4088 712 ? S 15:46 0:00 /usr/bin/runsv /marathon-lb/service/haproxy
root 9 0.1 0.1 142676 23496 ? Sl 15:46 0:00 python3 /marathon-lb/marathon_lb.py --syslog-socket /dev/null --haproxy-config /mar
root 10 0.1 0.0 20320 3128 ? S 15:46 0:00 /bin/bash ./run
root 1255 0.0 0.0 20332 3360 ? Ss+ 15:52 0:00 /bin/bash
root 1325 0.0 0.0 20332 3356 ? Ss 15:52 0:00 /bin/bash
root 1675 0.0 0.0 40504 10636 ? Ss 15:53 0:00 haproxy -p /tmp/haproxy.pid -f /marathon-lb/haproxy.cfg -D -sf 1519
root 1870 0.0 0.0 4224 704 ? S 15:53 0:00 sleep 0.5
root 1871 0.0 0.0 34492 2824 ? R+ 15:53 0:00 ps waux
root@ip-10-0-6-34:/marathon-lb# |
zombies seems related to mesos-executor, not marathon-lb, also had a host view :
from the namespace :
|
That's quite strange. What version of Mesos? I'm still not seeing the same thing: ip-10-0-6-34 ~ # ps waux | grep haproxy
root 17189 0.0 0.0 20256 3060 ? Ss 15:46 0:00 /bin/bash /marathon-lb/run sse -m http://master.mesos:8080 --health-check --haproxy-map --group external
root 17196 0.0 0.0 4088 712 ? S 15:46 0:00 /usr/bin/runsv /marathon-lb/service/haproxy
root 17197 0.0 0.1 144812 23736 ? Sl 15:46 0:01 python3 /marathon-lb/marathon_lb.py --syslog-socket /dev/null --haproxy-config /marathon-lb/haproxy.cfg --ssl-certs /etc/ssl/cert.pem --command sv reload /marathon-lb/service/haproxy --sse -m http://master.mesos:8080 --health-check --haproxy-map --group external
root 19815 0.1 0.0 40560 11764 ? Ss 15:53 0:09 haproxy -p /tmp/haproxy.pid -f /marathon-lb/haproxy.cfg -D -sf 1519
root 22953 0.0 0.0 4404 696 pts/0 S+ 18:10 0:00 grep --colour=auto haproxy
ip-10-0-6-34 ~ # |
mesos compiled from git master ( 1.1.0 ) , Just recompiled / tested a few minutes ago, same bug. mesos isolators : cni : simple loopback + bridge
marathon configuration : {
"id": "/internal/proxy/external",
"cmd": "/marathon-lb/run sse --marathon https://marathon:8443 --auth-credentials user:pass --group 'external' --ssl-certs /certs --max-serv-port-ip-per-task 20050",
"cpus": 0.01,
"mem": 128,
"disk": 0,
"instances": 2,
"container": {
"type": "MESOS",
"volumes": [
{
"containerPath": "/certs",
"hostPath": "/config/haproxy/certs",
"mode": "RO"
},
{
"containerPath": "/marathon-lb/templates",
"hostPath": "/config/haproxy/templates",
"mode": "RO"
}
],
"docker": {
"image": "mesosphere/marathon-lb:latest",
"forcePullImage": true
}
},
"env": {
"PORTS": "9090"
},
"healthChecks": [
{
"path": "/_haproxy_health_check",
"protocol": "HTTP",
"gracePeriodSeconds": 10,
"intervalSeconds": 10,
"timeoutSeconds": 2,
"maxConsecutiveFailures": 3,
"ignoreHttp1xx": false,
"port": 9090
}
],
"portDefinitions": [],
"ipAddress": {
"groups": [],
"labels": {},
"discovery": {
"ports": [
{
"number": 9090,
"name": "admin",
"protocol": "tcp",
"labels": {}
}
]
},
"networkName": "vlan"
}
} |
I think it's worth filing an issue over at https://issues.apache.org/jira/secure/Dashboard.jspa. I suspect this is related to Mesos, rather than MLB specifically. |
I'm going to close this for now, as I suspect it's a core Mesos issue. |
How many time it need to remove old process ?
|
Thanks @vixns |
I'm using latest docker image and mesos containerizer.
On each topology change, the old haproxy process become a zombie.
The text was updated successfully, but these errors were encountered: