-
Notifications
You must be signed in to change notification settings - Fork 1.8k
New issue
Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.
By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.
Already on GitHub? Sign in to your account
spl_kmem_cache() spinning #1454
Comments
Did your system have no swap, swap on a zvol or swap on a non-zvol? If this occurred when using swap on a zvol, ryao/spl@5717902 might fix this. |
@ryao This occurred on the MDS in a Lustre+ZFS configuration. The node was running diskless and there was swap device on the system (zvol or otherwise). I suspect Lustre may be to blame here by progressively consuming more memory and never providing a shrinker hook to release it. But I haven't investigate yet. Has your ryao/spl@5717902 patch been observed to improve things on systems with zvol swap? I can see how it might, but as you point out there also a distinct possibility the system will stall. |
I did not control variables tightly enough when testing ryao/spl@5717902 earlier. It does not appear to have any effect. The additional issue that I spotted might be happening, but determining that requires writing a patch for it and doing more testing. |
@behlendorf If you can still reproduce this, it would be helpful to profile this using perf and generate a flame graph like I described here: |
@behlendorf I wrote some notes on how to use flame graphs on the Gentoo Wiki: https://wiki.gentoo.org/wiki/ZFSOnLinux_Development_Guide#Flame_Graphs |
I've been unable to reproduce this issue with the latest code. I'm going to close it out. |
Observed while running mds-survey.
The text was updated successfully, but these errors were encountered: