-
Notifications
You must be signed in to change notification settings - Fork 2.4k
New issue
Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.
By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.
Already on GitHub? Sign in to your account
pod resource limits: error creating cgroup path: subtree_control: ENOENT #15074
Comments
pod resource limits test is in code that @cdoern just merged last week:
|
sdnotify test is systemd, so @vrothberg might be the best person to look at it, but it also could be crun, so, ping, @giuseppe also:
Lots more permission and SELinux errors, make me strongly suspect that SELinux is broken on these systems. It might be that the only way to debug is to ssh into one of them. |
@lsm5 hint for next time: file the issue first, then go to the broken PR and find links to all the failing logs, paste them in the issue, and then resubmit the PR with |
the only reason this should fail is if arm does not have subtree control which I find highly unlikely. the subtree_control file is less related to my resource limits work and more related to cgroup creation in general. I know where this is done in containers/common but still... an issue like this makes me think the kernel is missing some things when complied. |
could also be |
True @giuseppe but libpod_parent is created (if it does not exist) before subtree control I believe? |
then |
It's v2. I'm doing the Cirrus rerun-with-terminal thing, and trying to reproduce it, and can't: |
Still failing, but @lsm5 believes it might be a flake (which is consistent with my findings in the rerun terminal). I don't know if that's better or worse. |
I'll be darned. It is a flake. |
A friendly reminder that this issue had no activity for 30 days. |
|
Background: in order to add aarch64 tests, we had to add emergency skips to a lot of failing tests. No attempt was ever made to understand why they were failing. Fast forward to today, I filed containers#15888 just to see if tests are still failing. Looks like a number of them are fixed. (Yes, magically). Remove those skips. See: containers#15074, containers#15277 Signed-off-by: Ed Santiago <[email protected]>
Still happening on f38:
|
Seen just now on my RH laptop:
Passed on rerun. Again, this is my RH laptop, not aarch64. |
- containers#15074 ("subtree_control" flake). The flake is NOT FIXED, I saw it six months ago on my (non-aarch64) laptop. However, it looks like the frequent-flake-on-aarch64 bug is resolved. I've been testing in containers#17831 and have not seen it. So, tentatively remove the skip and see what happens. - Closes: containers#19407 (broken tar, "duplicates of file paths") All Fedoras now have a fixed tar. Debian DOES NOT, but we're handling that in our build-ci-vm code. I.e., the Debian VM we're using has a working tar even though there's currently a broken tar out in the wild. Added distro-integration tag so we can catch future problems like this in OpenQA. - Closes: containers#19471 (brq / blkio / loopbackfs in rawhide) Bug appears to be fixed in rawhide, at least in the VMs we're using now. Added distro-integration tag because this test obviously relies on other system stuff. Signed-off-by: Ed Santiago <[email protected]>
Seen after a long absence, f40 root, in parallel system tests though I doubt the parallel has anything to do with anything. |
Ping, seeing this one often in parallel system tests.
|
Continuing to see this often in parallel system tests
|
adding some code through containers/common#2158 to help debugging this issue |
Is this a BUG REPORT or FEATURE REQUEST? (leave only one on its own line)
/kind bug
Description
aarch64 CI enablement at #14801 is experiencing failures in the system tests. This issue is a placeholder for tracking and using in FIXME comments for
skip_if_aarch64
.The text was updated successfully, but these errors were encountered: