-
Notifications
You must be signed in to change notification settings - Fork 1.8k
New issue
Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.
By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.
Already on GitHub? Sign in to your account
many errors "attempt to access beyond end of device" after upgrade to 0.7.10 #7906
Comments
I've just got bitten by a similar issue. Upon updating CentOS 7.5 kmod-zfs-0.7.9-1.el7_4.x86_64 to 0.7.10-1.el7_4.x86_64. After the reboot members of one vdev got degraded, with one of them eventually getting faulted (the disks are HUS724040ALS640, 512k, but the pool is created with ashift=12):
/var/log/messages is flooded with messages of write attempts beyond the partition boundaries:
Comparing the affected two disks to the other vdev members shows decreased zfs partition size:
vs good one:
The affected disks show no errors in the grown defect list. |
@tonyhutter will be putting out an 0.7.11 release shortly which reverts the change. For the moment, you'll want to rollback to 0.7.9 on systems experiencing this issue. |
Closing, 0.7.11 was released with the fix, https://github.com/zfsonlinux/zfs/releases/tag/zfs-0.7.11 |
thanks for the quick fix! |
0.7.10 will not be committed due to a regression: openzfs/zfs#7906 Package-Manager: Portage-2.3.40, Repoman-2.3.9
Between openzfs/zfs#7909,openzfs/zfs#7899, openzfs/zfs#7906 and http://list.zfsonlinux.org/pipermail/zfs-discuss/2018-September/032318.html, it seems like 0.7.10 should be clearly marked as "bad".
0.7.10 will not be committed due to a regression: openzfs/zfs#7906 Package-Manager: Portage-2.3.40, Repoman-2.3.9
System information
Describe the problem you're observing
ZFS is trying to access sectors beyond the end of device, causing the disk to be marked as failed and the pool to be degraded or unavailable after upgrading from 0.7.9 to 0.7.10.
It looks related to #7724, for which (if I read the commit logs right) the fix landed in 0.7.10.
Describe how to reproduce the problem
Upgrade from 0.7.9 to 0.7.10. The problem is observed only on my pool with 4Kn drives (pool1), and not on pool2 which has drives with 512 byte sectors. Both pools were created with ZFS 0.7.9 and I gave the full disks to ZFS.
When I rolled back to 0.7.9, everything worked fine again.
Include any warning/errors/backtraces from the system logs
I'm happy to provide any additional debug info.
The text was updated successfully, but these errors were encountered: