-
Notifications
You must be signed in to change notification settings - Fork 2.4k
New issue
Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.
By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.
Already on GitHub? Sign in to your account
lstat /sys/fs/cgroup/devices/machine.slice/libpod-SHA.scope: ENOENT #11784
Comments
@giuseppe PTAL |
runc generates that error. Not sure if it is a regression but it appeared with: opencontainers/runc@cbb0a79 Simple reproducer (as root):
@kolyshkin FYI |
A friendly reminder that this issue had no activity for 30 days. |
Since this is not a Podman issue, should I close this? |
A friendly reminder that this issue had no activity for 30 days. |
Still happening.
Please remember that these do not cause actual CI failures, so my flake logger only catches them if they're present in an actual CI-failure-causing flake. The stats above are probably an underrepresentation. |
There's a race condition (see below). It seems that the freezer state has changed when attempting to freeze (i.e., it shouldn't attempt to freeze afaiks).
I used the following diff to get the error log: diff --git a/libcontainer/cgroups/fs2/freezer.go b/libcontainer/cgroups/fs2/freezer.go
index 8917a6411d68..b3ed1626c851 100644
--- a/libcontainer/cgroups/fs2/freezer.go
+++ b/libcontainer/cgroups/fs2/freezer.go
@@ -12,9 +12,11 @@ import (
"github.com/opencontainers/runc/libcontainer/cgroups"
"github.com/opencontainers/runc/libcontainer/configs"
+ "github.com/sirupsen/logrus"
)
func setFreezer(dirPath string, state configs.FreezerState) error {
+ logrus.Errorf("STATE: %s", state)
var stateStr string
switch state {
case configs.Undefined: @kolyshkin @cyphar could you have a look? I am not familiar with the runc code and think you know where to poke. |
From what I see, this is just two calls to I will take a closer look later. |
Thanks! To elaborate on why I think there's a race. Each time it fails, state is |
A friendly reminder that this issue had no activity for 30 days. |
@kolyshkin Any progress on this? |
Had no time to look at it, hopefully later this week (I have a separate browser window opened with this as a reminder 😁 ) |
A friendly reminder that this issue had no activity for 30 days. |
Not stale |
I think the cgroup could have been cleaned up by systemd while runc is trying to use it. Should we close this issue? I don't think there is anything we can do from the Podman side |
I agree. |
To silence my find-obsolete-skips script: - containers#11784 : issue closed wont-fix - containers#15013 : issue closed, we no longer test with runc - containers#15014 : bump timeout, see if that fixes things - containers#15025 : issue closed, we no longer test with runc ...and one FIXME not associated with an issue, ubuntu-related, and we no longer test ubuntu. Signed-off-by: Ed Santiago <[email protected]>
To silence my find-obsolete-skips script, remove the '#' from the following issues in skip messages: containers#11784 containers#15013 containers#15025 containers#17433 containers#17436 containers#17456 Also update the messages to reflect the fact that the issues will never be fixed. Also remove ubuntu skips: we no longer test ubuntu. Also remove one buildah skip that is no longer applicable: Fixes: containers#17520 Signed-off-by: Ed Santiago <[email protected]>
New flake in f33-root:
This is not (as of this writing) a flake that will cause a CI failure, because (as of this writing) system tests and integration tests do not check for extra cruft. THIS IS GOING TO CHANGE, at least in system tests.
I cannot reproduce with podman-3.4.0-0.10.rc2.fc33, in 30 minutes of looping, but my
cirrus-flake-grep
tool shows this happening as far back as June (when I started collecting CI logs). All the instances I see are root; none rootless.The text was updated successfully, but these errors were encountered: