Skip to content
This repository has been archived by the owner on Feb 26, 2020. It is now read-only.

SPL stack trace during scrub (rc11) #181

Closed
jpfingstmann opened this issue Oct 15, 2012 · 2 comments
Closed

SPL stack trace during scrub (rc11) #181

jpfingstmann opened this issue Oct 15, 2012 · 2 comments

Comments

@jpfingstmann
Copy link

When running a scrub, I'm sometimes getting the following stack trace (happens always sometime during a scrub, multiple times, just not always immediately at start):

[2124360.137047] SPL: Showing stack for process 16930
[2124360.137049] Pid: 16930, comm: txg_sync Tainted: P O 3.4.5-gentoo #2
[2124360.137051] Call Trace:
[2124360.137056] [] spl_debug_dumpstack+0x2b/0x2d [spl]
[2124360.137062] [] sanitize_flags+0x73/0x84 [spl]
[2124360.137068] [] kmalloc_nofail+0x21/0x3e [spl]
[2124360.137075] [] kmem_alloc_debug+0x160/0x2ce [spl]
[2124360.137081] [] ? kmalloc_nofail+0x2c/0x3e [spl]
[2124360.137101] [] ? ddt_zap_prefetch+0x10/0x10 [zfs]
[2124360.137120] [] ddt_zap_lookup+0x39/0xbc [zfs]
[2124360.137128] [] ? __gethrtime+0x11/0x1f [spl]
[2124360.137147] [] ddt_class_contains+0x9d/0xd8 [zfs]
[2124360.137175] [] dsl_scan_visitbp.isra.3+0x97a/0x984 [zfs]
[2124360.137189] [] ? arc_access.isra.6+0x15c/0x182 [zfs]
[2124360.137218] [] dsl_scan_visitbp.isra.3+0x545/0x984 [zfs]
[2124360.137233] [] ? arc_access.isra.6+0x15c/0x182 [zfs]
[2124360.137261] [] dsl_scan_visitbp.isra.3+0x36a/0x984 [zfs]
[2124360.137276] [] ? arc_access.isra.6+0x15c/0x182 [zfs]
[2124360.137304] [] dsl_scan_visitbp.isra.3+0x36a/0x984 [zfs]
[2124360.137319] [] ? arc_access.isra.6+0x15c/0x182 [zfs]
[2124360.137348] [] dsl_scan_visitbp.isra.3+0x36a/0x984 [zfs]
[2124360.137362] [] ? arc_access.isra.6+0x15c/0x182 [zfs]
[2124360.137391] [] dsl_scan_visitbp.isra.3+0x36a/0x984 [zfs]
[2124360.137405] [] ? arc_access.isra.6+0x15c/0x182 [zfs]
[2124360.137433] [] dsl_scan_visitbp.isra.3+0x36a/0x984 [zfs]
[2124360.137448] [] ? arc_access.isra.6+0x114/0x182 [zfs]
[2124360.137477] [] dsl_scan_visitbp.isra.3+0x36a/0x984 [zfs]
[2124360.137491] [] ? arc_access.isra.6+0x114/0x182 [zfs]
[2124360.137508] [] ? arc_read_nolock+0x1de/0x653 [zfs]
[2124360.137545] [] dsl_scan_visitbp.isra.3+0x68c/0x984 [zfs]
[2124360.137574] [] dsl_scan_visitds+0x171/0x3fb [zfs]
[2124360.137603] [] dsl_scan_sync+0x486/0x7d8 [zfs]
[2124360.137634] [] spa_sync+0x537/0x86d [zfs]
[2124360.137667] [] txg_sync_thread+0x27f/0x3ba [zfs]
[2124360.137701] [] ? txg_thread_wait.isra.2+0x2b/0x2b [zfs]
[2124360.137708] [] thread_generic_wrapper+0x6c/0x79 [spl]
[2124360.137715] [] ? __thread_create+0x2c2/0x2c2 [spl]
[2124360.137719] [] kthread+0x86/0x8e
[2124360.137722] [] kernel_thread_helper+0x4/0x10
[2124360.137725] [] ? retint_restore_args+0x5/0x6
[2124360.137728] [] ? gs_change+0x13/0x13

I did apply chrisrd/zfs@98f1e44 in an attempt to fix this, but it didn't help (it was part of a larger patchset that seemed to already be applied to rc11).

I'll keep my installation as-is for troubleshooting, let me know what I can do to help.

FYI - I'm running 64-bit gentoo with an Intel E5606 w/ 24 GB of ram (10 allocated to Dom0, I'm running a Xen kernel - I can revert back to plain Linux if needed).

Pool info:
pool: datastorage
state: ONLINE
scan: scrub repaired 0 in 37h58m with 0 errors on Sun Oct 14 14:31:47 2012
config:

    NAME                                               STATE     READ WRITE CKSUM
    datastorage                                        ONLINE       0     0     0
      raidz2-0                                         ONLINE       0     0     0
        pci-0000:01:00.0-sas-0x4433221106000000-lun-0  ONLINE       0     0     0
        pci-0000:01:00.0-sas-0x4433221104000000-lun-0  ONLINE       0     0     0
        pci-0000:01:00.0-sas-0x4433221105000000-lun-0  ONLINE       0     0     0
        pci-0000:01:00.0-sas-0x4433221107000000-lun-0  ONLINE       0     0     0
        pci-0000:01:00.0-sas-0x4433221102000000-lun-0  ONLINE       0     0     0
        pci-0000:01:00.0-sas-0x4433221101000000-lun-0  ONLINE       0     0     0
        pci-0000:01:00.0-sas-0x4433221100000000-lun-0  ONLINE       0     0     0
        pci-0000:01:00.0-sas-0x4433221103000000-lun-0  ONLINE       0     0     0
        pci-0000:04:00.0-sas-0x4433221105000000-lun-0  ONLINE       0     0     0
        pci-0000:04:00.0-sas-0x4433221104000000-lun-0  ONLINE       0     0     0

zfs list:
NAME USED AVAIL REFER MOUNTPOINT
datastorage 7.64T 6.58T 4.66T /datastorage
datastorage/Users 722G 6.58T 55.0K /datastorage/Users
datastorage/Users/jpfingstmann 694G 6.58T 445G /datastorage/Users/jpfingstmann
datastorage/Users/jpfingstmann/pcbackup 16.9G 6.58T 16.9G /datastorage/Users/jpfingstmann/pcbackup
datastorage/Users/kiba 27.8G 1.93T 27.8G /datastorage/Users/kiba
datastorage/crypt.zfs 3.31G 6.58T 224M -
datastorage/xentest-1 10.3G 6.58T 7.74G -

@behlendorf
Copy link
Contributor

These stacks were for debugging and have been fixed in master. You can grab the last source or wait a few weeks for -rc12.

@behlendorf
Copy link
Contributor

...or cherry pick the following commits in to ZFS

openzfs/zfs@6f53a6a Switch KM_SLEEP to KM_PUSHPAGE
openzfs/zfs@beb9994 Switch KM_SLEEP to KM_PUSHPAGE
openzfs/zfs@d75d6f2 Switch KM_SLEEP to KM_PUSHPAGE
openzfs/zfs@95fd8c9 Switch KM_SLEEP to KM_PUSHPAGE

Sign up for free to subscribe to this conversation on GitHub. Already have an account? Sign in.
Labels
None yet
Projects
None yet
Development

No branches or pull requests

2 participants