I got the following panic on 10.0-CURRENT from two days ago while receiving an incremental snapshot to a certain pool: (kgdb) where #0 doadump (textdump=0) at pcpu.h:229 #1 0xffffffff8031a3ce in db_dump (dummy=<value optimized out>, dummy2=0, dummy3=0, dummy4=0x0) at /usr/src/sys/ddb/db_command.c:543 #2 0xffffffff80319eca in db_command (last_cmdp=<value optimized out>, cmd_table=<value optimized out>, dopager=1) at /usr/src/sys/ddb/db_command.c:449 #3 0xffffffff80319c82 in db_command_loop () at /usr/src/sys/ddb/db_command.c:502 #4 0xffffffff8031c5d0 in db_trap (type=<value optimized out>, code=0) at /usr/src/sys/ddb/db_main.c:231 #5 0xffffffff805d0da3 in kdb_trap (type=3, code=0, tf=<value optimized out>) at /usr/src/sys/kern/subr_kdb.c:654 #6 0xffffffff8087fdc3 in trap (frame=0xffffff80dc9d6520) at /usr/src/sys/amd64/amd64/trap.c:579 #7 0xffffffff80869cb2 in calltrap () at exception.S:228 #8 0xffffffff805d058e in kdb_enter (why=0xffffffff80a47e7a "panic", msg=<value optimized out>) at cpufunc.h:63 #9 0xffffffff80599216 in panic (fmt=<value optimized out>) at /usr/src/sys/kern/kern_shutdown.c:747 #10 0xffffffff8130323f in assfail3 (a=<value optimized out>, lv=<value optimized out>, op=<value optimized out>, rv=<value optimized out>, f=<value optimized out>, l=<value optimized out>) at /usr/src/sys/modules/opensolaris/../../cddl/compat/opensolaris/kern/opensolaris_cmn_err.c:89 #11 0xffffffff8117924e in zfs_space_delta_cb (bonustype=<value optimized out>, data=0xffffff8015eeb8c0, userp=0xfffffe004261c640, groupp=0xfffffe004261c648) at /usr/src/sys/modules/zfs/../../cddl/contrib/opensolaris/uts/common/fs/zfs/zfs_vfsops.c:625 #12 0xffffffff8110003b in dmu_objset_userquota_get_ids (dn=0xfffffe004261c358, before=0, tx=<value optimized out>) at /usr/src/sys/modules/zfs/../../cddl/contrib/opensolaris/uts/common/fs/zfs/dmu_objset.c:1249 #13 0xffffffff811071b6 in dnode_sync (dn=0xfffffe004261c358, tx=0xfffffe00186e1300) at /usr/src/sys/modules/zfs/../../cddl/contrib/opensolaris/uts/common/fs/zfs/dnode_sync.c:554 #14 0xffffffff810ff98b in dmu_objset_sync_dnodes (list=0xfffffe00691a5250, newlist=<value optimized out>, tx=<value optimized out>) at /usr/src/sys/modules/zfs/../../cddl/contrib/opensolaris/uts/common/fs/zfs/dmu_objset.c:910 #15 0xffffffff810ff825 in dmu_objset_sync (os=0xfffffe00691a5000, pio=<value optimized out>, tx=0xfffffe00186e1300) at /usr/src/sys/modules/zfs/../../cddl/contrib/opensolaris/uts/common/fs/zfs/dmu_objset.c:1027 #16 0xffffffff8110cb0d in dsl_dataset_sync (ds=0xfffffe001f3d0c00, zio=0x780, tx=0xfffffe00186e1300) at /usr/src/sys/modules/zfs/../../cddl/contrib/opensolaris/uts/common/fs/zfs/dsl_dataset.c:1411 #17 0xffffffff8111399a in dsl_pool_sync (dp=0xfffffe0069ec4000, txg=<value optimized out>) at /usr/src/sys/modules/zfs/../../cddl/contrib/opensolaris/uts/common/fs/zfs/dsl_pool.c:409 #18 0xffffffff8112f0ee in spa_sync (spa=0xfffffe0050f00000, txg=3292) at /usr/src/sys/modules/zfs/../../cddl/contrib/opensolaris/uts/common/fs/zfs/spa.c:6328 #19 0xffffffff81137c45 in txg_sync_thread (arg=0xfffffe0069ec4000) at /usr/src/sys/modules/zfs/../../cddl/contrib/opensolaris/uts/common/fs/zfs/txg.c:493 #20 0xffffffff80569c1a in fork_exit (callout=0xffffffff811378d0 <txg_sync_thread>, arg=0xfffffe0069ec4000, frame=0xffffff80dc9d6c00) at /usr/src/sys/kern/kern_fork.c:991 #21 0xffffffff8086a1ee in fork_trampoline () at exception.S:602 #22 0x0000000000000000 in ?? () Current language: auto; currently minimal (kgdb) f 12 #12 0xffffffff8110003b in dmu_objset_userquota_get_ids (dn=0xfffffe004261c358, before=0, tx=<value optimized out>) at /usr/src/sys/modules/zfs/../../cddl/contrib/opensolaris/uts/common/fs/zfs/dmu_objset.c:1249 1249 error = used_cbs[os->os_phys->os_type](dn->dn_bonustype, data, (kgdb) p *dn $1 = {dn_struct_rwlock = {lock_object = {lo_name = 0xffffffff811da0a9 "dn->dn_struct_rwlock", lo_flags = 40960000, lo_data = 0, lo_witness = 0x0}, sx_lock = 1}, dn_link = {list_next = 0xfffffe0042629020, list_prev = 0xfffffe00691a5360}, dn_objset = 0xfffffe00691a5000, dn_object = 55652, dn_dbuf = 0xfffffe00427ad0e0, dn_handle = 0xfffffe0069f70128, dn_phys = 0xffffff8015eeb800, dn_type = DMU_OT_PLAIN_FILE_CONTENTS, dn_bonuslen = 192, dn_bonustype = 44 ',', dn_nblkptr = 1 '\001', dn_checksum = 0 '\0', dn_compress = 0 '\0', dn_nlevels = 1 '\001', dn_indblkshift = 14 '\016', dn_datablkshift = 0 '\0', dn_moved = 0 '\0', dn_datablkszsec = 10, dn_datablksz = 5120, dn_maxblkid = 0, dn_next_nblkptr = "\000\000\000", dn_next_nlevels = "\000\000\000", dn_next_indblkshift = "\000\000\000", dn_next_bonustype = ",\000\000", dn_rm_spillblk = "\000\000\000", dn_next_bonuslen = {192, 0, 0, 0}, dn_next_blksz = {0, 0, 0, 0}, dn_dbufs_count = 0, dn_dirty_link = {{ list_next = 0xfffffe00691a51f0, list_prev = 0xfffffe0042628ab0}, {list_next = 0x0, list_prev = 0x0}, {list_next = 0x0, list_prev = 0x0}, {list_next = 0x0, list_prev = 0x0}}, dn_mtx = {lock_object = { lo_name = 0xffffffff811da0bf "dn->dn_mtx", lo_flags = 40960000, lo_data = 0, lo_witness = 0x0}, sx_lock = 1}, dn_dirty_records = {{list_size = 208, list_offset = 0, list_head = { list_next = 0xfffffe004261c470, list_prev = 0xfffffe004261c470}}, {list_size = 208, list_offset = 0, list_head = {list_next = 0xfffffe004261c490, list_prev = 0xfffffe004261c490}}, {list_size = 208, list_offset = 0, list_head = {list_next = 0xfffffe004261c4b0, list_prev = 0xfffffe004261c4b0}}, {list_size = 208, list_offset = 0, list_head = {list_next = 0xfffffe004261c4d0, list_prev = 0xfffffe004261c4d0}}}, dn_ranges = {{avl_root = 0x0, avl_compar = 0xffffffff81106ec0 <free_range_compar>, avl_offset = 0, avl_numnodes = 0, avl_size = 40}, {avl_root = 0x0, avl_compar = 0xffffffff81106ec0 <free_range_compar>, avl_offset = 0, avl_numnodes = 0, avl_size = 40}, {avl_root = 0x0, avl_compar = 0xffffffff81106ec0 <free_range_compar>, avl_offset = 0, avl_numnodes = 0, avl_size = 40}, {avl_root = 0x0, avl_compar = 0xffffffff81106ec0 <free_range_compar>, avl_offset = 0, avl_numnodes = 0, avl_size = 40}}, dn_allocated_txg = 3292, dn_free_txg = 0, dn_assigned_txg = 0, dn_notxholds = {cv_description = 0xffffffff811da0dd "dn->dn_notxholds", cv_waiters = 0}, dn_dirtyctx = DN_UNDIRTIED, dn_dirtyctx_firstset = 0x0, dn_tx_holds = {rc_count = 0}, dn_holds = {rc_count = 3}, dn_dbufs_mtx = {lock_object = {lo_name = 0xffffffff811da0cb "dn->dn_dbufs_mtx", lo_flags = 40960000, lo_data = 0, lo_witness = 0x0}, sx_lock = 1}, dn_dbufs = {list_size = 224, list_offset = 176, list_head = {list_next = 0xfffffe004261c5f8, list_prev = 0xfffffe004261c5f8}}, dn_bonus = 0x0, dn_have_spill = 0, dn_zio = 0xfffffe00695af000, dn_oldused = 2560, dn_oldflags = 3, dn_olduid = 1001, dn_oldgid = 1001, dn_newuid = 0, dn_newgid = 0, dn_id_flags = 5, dn_zfetch = {zf_rwlock = {lock_object = {lo_name = 0xffffffff811dd156 "zf->zf_rwlock", lo_flags = 40960000, lo_data = 0, lo_witness = 0x0}, sx_lock = 1}, zf_stream = {list_size = 112, list_offset = 88, list_head = {list_next = 0xfffffe004261c688, list_prev = 0xfffffe004261c688}}, zf_dnode = 0xfffffe004261c358, zf_stream_cnt = 0, zf_alloc_fail = 0}} The incremental was created with: zfs send -i _at_2013-03-28_14:21 tank/home/fk_at_2013-04-01_12:31 piped through mbuffer and received with: zfs receive -v -u -F rockbox/backup/r500/tank/home/fk Reading the incremental directly from a file triggers the panic as as well, but sometimes it takes more than one attempt. The offending sa_magic in the panic message is always the same. The receiving pool appears to be okay: fk_at_r500 ~ $zpool status rockbox pool: rockbox state: ONLINE status: Some supported features are not enabled on the pool. The pool can still be used, but some features are unavailable. action: Enable all features using 'zpool upgrade'. Once this is done, the pool may no longer be accessible by software that does not support the features. See zpool-features(7) for details. scan: scrub repaired 0 in 0h3m with 0 errors on Mon Apr 1 13:57:35 2013 config: NAME STATE READ WRITE CKSUM rockbox ONLINE 0 0 0 label/rockbox.eli ONLINE 0 0 0 errors: No known data errors The feature that isn't yet enabled is lz4 but after upgrading a copy of the pool the panic was still reproducible. On the receiving side gzip compression is enabled. Fabian
This archive was generated by hypermail 2.4.0 : Wed May 19 2021 - 11:40:36 UTC