ubuntu 21.10 Linux i5 5.13.0-20-generic #20-Ubuntu SMP Fri Oct 15 14:21:35 UTC 2021 x86_64 x86_64 x86_64 GNU/Linux
zfs-2.0.6-1ubuntu2 zfs-kmod-2.0.6-1ubuntu2 $ rm -rfv .steam ... '.steam/steam/config/htmlcache/Cache/5f67979416a221e5_0' supprimé '.steam/steam/config/htmlcache/Cache/2b3b061ac9d6b292_0' supprimé '.steam/steam/config/htmlcache/Cache/88ec9cdf0f2e7cf6_0' supprimé '.steam/steam/config/htmlcache/Cache/446e83c112a55833_0' supprimé '.steam/steam/config/htmlcache/Cache/25bce87ba6a10af5_0' supprimé '.steam/steam/config/htmlcache/Cache/a409ef32a0f5a1b3_0' supprimé '.steam/steam/config/htmlcache/Cache/2e8722be934b8d51_0' supprimé '.steam/steam/config/htmlcache/Cache/cb27d7e85cfb9396_0' supprimé '.steam/steam/config/htmlcache/Cache/f7bb287f03ab70bb_0' supprimé '.steam/steam/config/htmlcache/Cache/18b17be83cac58df_0' supprimé '.steam/steam/config/htmlcache/Cache/9f7e378b5b8fe6cf_0' supprimé '.steam/steam/config/htmlcache/Cache/3742398e7e6ac7aa_0' supprimé stall [ 549.052760] VERIFY(0 == sa_handle_get_from_db(zfsvfs->z_os, db, zp, SA_HDL_SHARED, &zp->z_sa_hdl)) failed [ 549.052763] PANIC at zfs_znode.c:339:zfs_znode_sa_init() [ 549.052765] Showing stack for process 18526 [ 549.052766] CPU: 0 PID: 18526 Comm: rm Tainted: P OE 5.13.0-20-generic #20-Ubuntu [ 549.052768] Hardware name: System manufacturer System Product Name/PRIME Z270-A, BIOS 0505 11/08/2016 [ 549.052769] Call Trace: [ 549.052772] show_stack+0x52/0x58 [ 549.052776] dump_stack+0x7d/0x9c [ 549.052781] spl_dumpstack+0x29/0x2b [spl] [ 549.052791] spl_panic+0xd4/0xfc [spl] [ 549.052799] ? queued_spin_unlock+0x9/0x10 [zfs] [ 549.052889] ? do_raw_spin_unlock+0x9/0x10 [zfs] [ 549.052944] ? __raw_spin_unlock+0x9/0x10 [zfs] [ 549.052998] ? dmu_buf_replace_user+0x65/0x80 [zfs] [ 549.053053] ? dmu_buf_set_user+0x13/0x20 [zfs] [ 549.053107] ? dmu_buf_set_user_ie+0x15/0x20 [zfs] [ 549.053160] zfs_znode_sa_init+0xd9/0xe0 [zfs] [ 549.053242] zfs_znode_alloc+0x101/0x580 [zfs] [ 549.053325] ? dmu_buf_unlock_parent+0x5d/0x90 [zfs] [ 549.053380] ? do_raw_spin_unlock+0x9/0x10 [zfs] [ 549.053436] ? dbuf_read_impl.constprop.0+0x30a/0x3e0 [zfs] [ 549.053489] ? dbuf_rele_and_unlock+0x13b/0x520 [zfs] [ 549.053541] ? __cond_resched+0x1a/0x50 [ 549.053544] ? __raw_callee_save___native_queued_spin_unlock+0x15/0x23 [ 549.053547] ? queued_spin_unlock+0x9/0x10 [zfs] [ 549.053597] ? do_raw_spin_unlock+0x9/0x10 [zfs] [ 549.053647] ? __cond_resched+0x1a/0x50 [ 549.053648] ? down_read+0x13/0x90 [ 549.053650] ? __raw_callee_save___native_queued_spin_unlock+0x15/0x23 [ 549.053652] ? queued_spin_unlock+0x9/0x10 [zfs] [ 549.053711] ? do_raw_spin_unlock+0x9/0x10 [zfs] [ 549.053770] ? __raw_callee_save___native_queued_spin_unlock+0x15/0x23 [ 549.053773] ? dmu_object_info_from_dnode+0x8e/0xa0 [zfs] [ 549.053829] zfs_zget+0x235/0x280 [zfs] [ 549.053909] zfs_dirent_lock+0x420/0x560 [zfs] [ 549.053990] zfs_dirlook+0x91/0x2d0 [zfs] [ 549.054070] zfs_lookup+0x257/0x400 [zfs] [ 549.054149] zpl_lookup+0xcb/0x220 [zfs] [ 549.054227] ? __d_alloc+0x138/0x1f0 [ 549.054229] __lookup_hash+0x70/0xa0 [ 549.054231] ? __cond_resched+0x1a/0x50 [ 549.054232] do_unlinkat+0x135/0x2e0 [ 549.054234] __x64_sys_unlinkat+0x3a/0x60 [ 549.054235] do_syscall_64+0x61/0xb0 [ 549.054237] ? do_syscall_64+0x6e/0xb0 [ 549.054238] ? do_syscall_64+0x6e/0xb0 [ 549.054239] ? __x64_sys_write+0x19/0x20 [ 549.054241] ? do_syscall_64+0x6e/0xb0 [ 549.054242] ? asm_sysvec_apic_timer_interrupt+0xa/0x20 [ 549.054243] entry_SYSCALL_64_after_hwframe+0x44/0xae [ 549.054246] RIP: 0033:0x7ff2c3aaf1fb [ 549.054248] Code: 73 01 c3 48 8b 0d 35 ec 0f 00 f7 d8 64 89 01 48 83 c8 ff c3 66 2e 0f 1f 84 00 00 00 00 00 90 f3 0f 1e fa b8 07 01 00 00 0f 05 <48> 3d 01 f0 ff ff 73 01 c3 48 8b 0d 05 ec 0f 00 f7 d8 64 89 01 48 [ 549.054249] RSP: 002b:00007ffc261242b8 EFLAGS: 00000246 ORIG_RAX: 0000000000000107 [ 549.054251] RAX: ffffffffffffffda RBX: 00005607c3b07310 RCX: 00007ff2c3aaf1fb [ 549.054252] RDX: 0000000000000000 RSI: 00005607c3b07418 RDI: 0000000000000008 [ 549.054253] RBP: 00005607c3aa2440 R08: 0000000000000003 R09: 0000000000000000 [ 549.054253] R10: 0000000000000000 R11: 0000000000000246 R12: 0000000000000000 [ 549.054254] R13: 00007ffc26124490 R14: 00007ffc26124490 R15: 0000000000000002 $ zpool scrub rpool pool: rpool state: ONLINE scan: scrub repaired 0B in 00:07:17 with 0 errors on Sun Oct 24 00:12:51 2021 config: NAME STATE READ WRITE CKSUM rpool ONLINE 0 0 0 0518af4d-aa46-5a45-a87f-3ac806eed4c1 ONLINE 0 0 0 errors: No known data errors -- You received this bug notification because you are a member of Kernel Packages, which is subscribed to zfs-linux in Ubuntu. https://bugs.launchpad.net/bugs/1906476 Title: PANIC at zfs_znode.c:335:zfs_znode_sa_init() // VERIFY(0 == sa_handle_get_from_db(zfsvfs->z_os, db, zp, SA_HDL_SHARED, &zp->z_sa_hdl)) failed Status in Native ZFS for Linux: New Status in linux package in Ubuntu: Invalid Status in ubuntu-release-upgrader package in Ubuntu: Confirmed Status in zfs-linux package in Ubuntu: Fix Released Status in linux source package in Impish: Fix Released Status in ubuntu-release-upgrader source package in Impish: Confirmed Status in zfs-linux source package in Impish: Fix Released Bug description: Since today while running Ubuntu 21.04 Hirsute I started getting a ZFS panic in the kernel log which was also hanging Disk I/O for all Chrome/Electron Apps. I have narrowed down a few important notes: - It does not happen with module version 0.8.4-1ubuntu11 built and included with 5.8.0-29-generic - It was happening when using zfs-dkms 0.8.4-1ubuntu16 built with DKMS on the same kernel and also on 5.8.18-acso (a custom kernel). - For whatever reason multiple Chrome/Electron apps were affected, specifically Discord, Chrome and Mattermost. In all cases they seem (but I was unable to strace the processes so it was a bit hard ot confirm 100% but by deduction from /proc/PID/fd and the hanging ls) they seem hung trying to open files in their 'Cache' directory, e.g. ~/.cache/google-chrome/Default/Cache and ~/.config/Mattermost/Cache .. while the issue was going on I could not list that directory either "ls" would just hang. - Once I removed zfs-dkms only to revert to the kernel built-in version it immediately worked without changing anything, removing files, etc. - It happened over multiple reboots and kernels every time, all my Chrome apps weren't working but for whatever reason nothing else seemed affected. - It would log a series of spl_panic dumps into kern.log that look like this: Dec 2 12:36:42 optane kernel: [ 72.857033] VERIFY(0 == sa_handle_get_from_db(zfsvfs->z_os, db, zp, SA_HDL_SHARED, &zp->z_sa_hdl)) failed Dec 2 12:36:42 optane kernel: [ 72.857036] PANIC at zfs_znode.c:335:zfs_znode_sa_init() I could only find one other google reference to this issue, with 2 other users reporting the same error but on 20.04 here: https://github.com/openzfs/zfs/issues/10971 - I was not experiencing the issue on 0.8.4-1ubuntu14 and fairly sure it was working on 0.8.4-1ubuntu15 but broken after upgrade to 0.8.4-1ubuntu16. I will reinstall those zfs-dkms versions to verify that. There were a few originating call stacks but the first one I hit was Call Trace: dump_stack+0x74/0x95 spl_dumpstack+0x29/0x2b [spl] spl_panic+0xd4/0xfc [spl] ? sa_cache_constructor+0x27/0x50 [zfs] ? _cond_resched+0x19/0x40 ? mutex_lock+0x12/0x40 ? dmu_buf_set_user_ie+0x54/0x80 [zfs] zfs_znode_sa_init+0xe0/0xf0 [zfs] zfs_znode_alloc+0x101/0x700 [zfs] ? arc_buf_fill+0x270/0xd30 [zfs] ? __cv_init+0x42/0x60 [spl] ? dnode_cons+0x28f/0x2a0 [zfs] ? _cond_resched+0x19/0x40 ? _cond_resched+0x19/0x40 ? mutex_lock+0x12/0x40 ? aggsum_add+0x153/0x170 [zfs] ? spl_kmem_alloc_impl+0xd8/0x110 [spl] ? arc_space_consume+0x54/0xe0 [zfs] ? dbuf_read+0x4a0/0xb50 [zfs] ? _cond_resched+0x19/0x40 ? mutex_lock+0x12/0x40 ? dnode_rele_and_unlock+0x5a/0xc0 [zfs] ? _cond_resched+0x19/0x40 ? mutex_lock+0x12/0x40 ? dmu_object_info_from_dnode+0x84/0xb0 [zfs] zfs_zget+0x1c3/0x270 [zfs] ? dmu_buf_rele+0x3a/0x40 [zfs] zfs_dirent_lock+0x349/0x680 [zfs] zfs_dirlook+0x90/0x2a0 [zfs] ? zfs_zaccess+0x10c/0x480 [zfs] zfs_lookup+0x202/0x3b0 [zfs] zpl_lookup+0xca/0x1e0 [zfs] path_openat+0x6a2/0xfe0 do_filp_open+0x9b/0x110 ? __check_object_size+0xdb/0x1b0 ? __alloc_fd+0x46/0x170 do_sys_openat2+0x217/0x2d0 ? do_sys_openat2+0x217/0x2d0 do_sys_open+0x59/0x80 __x64_sys_openat+0x20/0x30 To manage notifications about this bug go to: https://bugs.launchpad.net/zfs/+bug/1906476/+subscriptions -- Mailing list: https://launchpad.net/~kernel-packages Post to : kernel-packages@lists.launchpad.net Unsubscribe : https://launchpad.net/~kernel-packages More help : https://help.launchpad.net/ListHelp