-
Notifications
You must be signed in to change notification settings - Fork 814
New issue
Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.
By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.
Already on GitHub? Sign in to your account
Add liveness probe #225
Add liveness probe #225
Conversation
[APPROVALNOTIFIER] This PR is APPROVED This pull-request has been approved by: leakingtapan The full list of commands accepted by this bot can be found here. The pull request process is described here
Needs approval from an approver in each of these files:
Approvers can indicate their approval by writing |
deploy/kubernetes/controller.yaml
Outdated
initialDelaySeconds: 10 | ||
timeoutSeconds: 3 | ||
periodSeconds: 2 | ||
failureThreshold: 1 |
There was a problem hiding this comment.
Choose a reason for hiding this comment
The reason will be displayed to describe this comment to others. Learn more.
failureThreshold: 1
might be too aggressive, in their example they actually have 2:
failureThreshold: 5
...
failureThreshold: 1
There was a problem hiding this comment.
Choose a reason for hiding this comment
The reason will be displayed to describe this comment to others. Learn more.
Looks like the test failures could have been caused because the the node driver restarted:
Logging pods the kubelet thinks is on node ip-172-20-50-17.us-west-2.compute.internal
Feb 26 02:54:34.076: INFO: kube-proxy-ip-172-20-50-17.us-west-2.compute.internal started at <nil> (0+0 container statuses recorded)
Feb 26 02:54:34.076: INFO: ebs-csi-node-xn2c8 started at 2019-02-26 02:33:06 +0000 UTC (0+3 container statuses recorded)
Feb 26 02:54:34.076: INFO: Container ebs-plugin ready: true, restart count 6
All 4 failures were on the volumes that take a long time to format and even if the formatting is attempted a second time, the 15 min timeout is not enough time.
There was a problem hiding this comment.
Choose a reason for hiding this comment
The reason will be displayed to describe this comment to others. Learn more.
Notice the driver keeps being restarted many time even without any traffic due to probe timeout. Update the periodSeconds
to 10s for less frequent probe, failureThreshold
to 5 for higher failure tolerance. With the change, driver is back stable.
da040ad
to
11535a4
Compare
11535a4
to
f80776c
Compare
/lgtm |
…t/cherry-pick-224-to-release-4.13 [release-4.13] OCPBUGS-13811: Volume unmount repeats after successful unmount, preventing pod delete
Is this a bug fix or adding new feature?
Fixes: #159
What is this PR about? / Why do we need it?
What testing is done?