-
Notifications
You must be signed in to change notification settings - Fork 493
New issue
Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.
By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.
Already on GitHub? Sign in to your account
Operator cancels rollout of ds/deployment #1300
Comments
If that is the case, shouldn't the |
I don't think that's possible. Just tried and got this error:
I think the things that are "rollout"-able are limited to Deployments, DaemonSets, and StatefulSets. Not sure if there's a way to configure that. |
Anything that changes the collector pods spec in deployment will cause a restarts - e.g. probably adding a new annotation
|
Yes, doing a rollout on the deployment should do that, but the controller immediately replaces the annotations back to the way it was, which seemingly cancels the rollout |
I meant adding the annotation on the collector CR, the operator will update the deployment and controller will restart the pod. |
Oh I see. That's a good workaround. |
I think this needs to be tackled properly, not worked around. This is problematic, for example, when spot instances are taken down. We have seen pods on non-existent nodes, and rollouts stuck because of this. How would a fix look? Naively I'd say we just exclude the rollout annotation from reconciliation. Would be happy to work on it. |
@verejoel that would be great. Thank you! |
@verejoel any chance you'd be able to take a look at this for the current week? If not, i'm happy to take a look. |
Please go ahead, I don’t have time to look at this too deeply at the moment
…On Fri, 12 Jan 2024 at 20:29, Jacob Aronoff ***@***.***> wrote:
@verejoel <https://github.com/verejoel> im going to reassign this one, as
its something i just ran into. Let me know if you've already begun. If not
@Toaddyan <https://github.com/Toaddyan> take a look
—
Reply to this email directly, view it on GitHub
<#1300 (comment)>,
or unsubscribe
<https://github.com/notifications/unsubscribe-auth/ALE5VOMGJBJSPVZDTREH6YLYOGFJ7AVCNFSM6AAAAAASUOHEHSVHI2DSMVQWIX3LMV43OSLTON2WKQ3PNVWWK3TUHMYTQOBZHAZTQNZVHA>
.
You are receiving this because you were mentioned.Message ID:
***@***.***>
|
So I was working on trying to re-produce the issue:
From the above conversation, I'm interpreting that upon a rollout restart of a type of resource (daemonset), I should see the problem of a subset of pods restarting, where I SHOULD be seeing ALL of the pods restarting. In my recent run down of this, I'm not seeing the OP's stated behavior of restarts. am I missing anything particular here? |
I can't say for sure this is what's happening, but it seems like it to me.
kubectl rollout restart
restarts the workload by adding/modifying an annotation to the pod template spec in order to trigger a rollout. It seems like when it does this, the controller is immediately reconciling and removing the new annotation. I assume this is because it's replacing the annotations field instead of doing some kind of merge patch / SSA.This causes the OpenTelemetry Collector DaemonSet to restart maybe one pod, but the rollout stops after that.
The text was updated successfully, but these errors were encountered: