-
Notifications
You must be signed in to change notification settings - Fork 2.5k
New issue
Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.
By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.
Already on GitHub? Sign in to your account
[v4.0-rhel] Do not error on signalling a just-stopped container #14727
[v4.0-rhel] Do not error on signalling a just-stopped container #14727
Conversation
Previous PR containers#12394 tried to address this, but made a mistake: containers that have just exited do not move to the Exited state but rather the Stopped state - as such, the code would never have run (there is no way we start `podman kill`, and the container transitions to Exited while we are doing it - that requires holding the container lock, which Kill already does). Fix the code to check Stopped as well (we omit Exited entirely but it's a cheap check and our state logic could change in the future). Also, return an error, instead of exiting cleanly - the Kill failed, after all. ErrCtrStateInvalid is already handled by the sig-proxy logic so there won't be issues. [NO NEW TESTS NEEDED] This fixes a race that I cannot reproduce myself, and I have no idea how we'd repro in CI. Signed-off-by: Matthew Heon <[email protected]>
LGTM |
/approve |
[APPROVALNOTIFIER] This PR is APPROVED This pull-request has been approved by: openshift-cherrypick-robot, rhatdan The full list of commands accepted by this bot can be found here. The pull request process is described here
Needs approval from an approver in each of these files:
Approvers can indicate their approval by writing |
Lot of netavark test failures, looking... |
Potentially a network issue (on the Github or Cirrus end)? Or maybe that release just disappeared... |
Restarting one of the failures to see if it's a flake /lgtm In case it is |
Seems to fail consistently |
@lsm5 what's up with this PR? Do we need it? If it's for RHEL please link a BZ. |
@mheon do you know what's up with CI? |
Potentially trying to grab a bad version of Netavark for this branch? @cevich PTAL |
Looking...
|
...sigh, the netavark CI VM images got pruned 😭 The cirrus-cron jobs have been failing since June and nobody noticed because monitoring didn't exist. The monitoring part is fixed by containers/netavark#325 but it may be too late for me to recover the VM images, still looking... |
...well If I can't be smart all the time, at least I can be lucky: containers/netavark#337 <-- should fix this after it merges and CI runs on that branch. |
@cevich Does this need a rebase? |
Looking. |
This is referring to |
...curious, it appears a recent job run by cirrus-cron on the |
...looks like it's working now. Just needed to re-run the jobs (I just did). |
https://bugzilla.redhat.com/show_bug.cgi?id=2097049 targets v4.1.1 where it's fixed. Please reopen if I am mistaken. |
This is an automated cherry-pick of #14533
/assign lsm5