Colin, thanks for the link to https://github.com/openzfs/zfs/issues/9966 ; unfortunately I think that's a different problem, my meta use seems less drastic than the github issue
arc_prune 4 1859269059 arc_meta_used 4 3590932168 arc_meta_limit 4 94868305920 arc_dnode_limit 4 4294967296 arc_meta_max 4 79969748624 arc_meta_min 4 16777216 It looks like there's still 14898557296 bytes available for arc_meta. Do you know off-hand what the comment about the l2arc issue is? My l2arc is 1TB, well above 320 gigs. Here's where my stats are now (the system is not currently under any real load): $ cat /proc/spl/kstat/zfs/arcstats 13 1 0x01 96 4608 10492531405 1374854032619736 name type data hits 4 6820744874 misses 4 953303948 demand_data_hits 4 1163033450 demand_data_misses 4 112593263 demand_metadata_hits 4 5398565264 demand_metadata_misses 4 771209639 prefetch_data_hits 4 3978361 prefetch_data_misses 4 4454061 prefetch_metadata_hits 4 255167799 prefetch_metadata_misses 4 65046985 mru_hits 4 2596176358 mru_ghost_hits 4 7960783 mfu_hits 4 3972213379 mfu_ghost_hits 4 37066000 deleted 4 624581823 mutex_miss 4 66283716 access_skip 4 112 evict_skip 4 59660318004 evict_not_enough 4 531586443 evict_l2_cached 4 758952236544 evict_l2_eligible 4 6629993222656 evict_l2_ineligible 4 1055824726016 evict_l2_skip 4 4093 hash_elements 4 1382698 hash_elements_max 4 7705756 hash_collisions 4 101258453 hash_chains 4 54374 hash_chain_max 4 6 p 4 3403412095 c 4 9636886736 c_min 4 4221281536 c_max 4 126491074560 size 4 9501317832 compressed_size 4 6533731328 uncompressed_size 4 12946349568 overhead_size 4 1324591616 hdr_size 4 125140504 data_size 4 5910385664 metadata_size 4 1947937280 dbuf_size 4 313455312 dnode_size 4 801582080 bonus_size 4 304865408 anon_size 4 69632 anon_evictable_data 4 0 anon_evictable_metadata 4 0 mru_size 4 2874187264 mru_evictable_data 4 2820684800 mru_evictable_metadata 4 11149312 mru_ghost_size 4 1934149120 mru_ghost_evictable_data 4 254414848 mru_ghost_evictable_metadata 4 1679734272 mfu_size 4 4984066048 mfu_evictable_data 4 3089700864 mfu_evictable_metadata 4 293028864 mfu_ghost_size 4 32707072 mfu_ghost_evictable_data 4 32707072 mfu_ghost_evictable_metadata 4 0 l2_hits 4 16313870 l2_misses 4 936990042 l2_feeds 4 1342134 l2_rw_clash 4 1 l2_read_bytes 4 73933849600 l2_write_bytes 4 218946362880 l2_writes_sent 4 45529 l2_writes_done 4 45529 l2_writes_error 4 0 l2_writes_lock_retry 4 21 l2_evict_lock_retry 4 0 l2_evict_reading 4 0 l2_evict_l1cached 4 0 l2_free_on_write 4 878 l2_abort_lowmem 4 496 l2_cksum_bad 4 0 l2_io_error 4 0 l2_size 4 130902416896 l2_asize 4 98157307392 l2_hdr_size 4 97951584 memory_throttle_count 4 0 memory_direct_count 4 856012 memory_indirect_count 4 233680 memory_all_bytes 4 135081009152 memory_free_bytes 4 120346943488 memory_available_bytes 3 118236303360 arc_no_grow 4 0 arc_tempreserve 4 0 arc_loaned_bytes 4 0 arc_prune 4 1859269059 arc_meta_used 4 3590932168 arc_meta_limit 4 94868305920 arc_dnode_limit 4 4294967296 arc_meta_max 4 79969748624 arc_meta_min 4 16777216 sync_wait_for_async 4 608494 demand_hit_predictive_prefetch 4 5716590 arc_need_free 4 0 arc_sys_free 4 2110640768 Thanks -- You received this bug notification because you are a member of Kernel Packages, which is subscribed to zfs-linux in Ubuntu. https://bugs.launchpad.net/bugs/1814983 Title: zfs poor sustained read performance from ssd pool Status in Native ZFS for Linux: Unknown Status in zfs-linux package in Ubuntu: Incomplete Bug description: Hello, I'm seeing substantially slower read performance from an ssd pool than I expected. I have two pools on this computer; one ('fst') is four sata ssds, the other ('srv') is nine spinning metal drives. With a long-running ripgrep process on the fst pool, performance started out really good and grew to astonishingly good (iirc ~30kiops, as measured by zpool iostat -v 1). However after a few hours the performance has dropped to 30-40 iops. top reports an arc_reclaim and many arc_prune processes to be consuming most of the CPU time. I've included a screenshot of top, some output from zpool iostat -v 1, and arc_summary, with "===" to indicate the start of the next command's output: === top (memory in gigabytes): top - 16:27:53 up 70 days, 16:03, 3 users, load average: 35.67, 35.81, 35.58 Tasks: 809 total, 19 running, 612 sleeping, 0 stopped, 0 zombie %Cpu(s): 0.0 us, 58.1 sy, 0.0 ni, 39.2 id, 2.6 wa, 0.0 hi, 0.0 si, 0.0 st GiB Mem : 125.805 total, 0.620 free, 96.942 used, 28.243 buff/cache GiB Swap: 5.694 total, 5.688 free, 0.006 used. 27.840 avail Mem PID USER PR NI VIRT RES SHR S %CPU %MEM TIME+ COMMAND 1523 root 20 0 0.0m 0.0m 0.0m R 100.0 0.0 290:52.26 arc_reclaim 4484 root 20 0 0.0m 0.0m 0.0m R 56.2 0.0 1:18.79 arc_prune 6225 root 20 0 0.0m 0.0m 0.0m R 56.2 0.0 1:11.92 arc_prune 7601 root 20 0 0.0m 0.0m 0.0m S 56.2 0.0 2:50.25 arc_prune 30891 root 20 0 0.0m 0.0m 0.0m S 56.2 0.0 1:33.08 arc_prune 3057 root 20 0 0.0m 0.0m 0.0m S 55.9 0.0 9:00.95 arc_prune 3259 root 20 0 0.0m 0.0m 0.0m R 55.9 0.0 3:16.84 arc_prune 24008 root 20 0 0.0m 0.0m 0.0m S 55.9 0.0 1:55.71 arc_prune 1285 root 20 0 0.0m 0.0m 0.0m R 55.6 0.0 3:20.52 arc_prune 5345 root 20 0 0.0m 0.0m 0.0m R 55.6 0.0 1:15.99 arc_prune 30121 root 20 0 0.0m 0.0m 0.0m S 55.6 0.0 1:35.50 arc_prune 31192 root 20 0 0.0m 0.0m 0.0m S 55.6 0.0 6:17.16 arc_prune 32287 root 20 0 0.0m 0.0m 0.0m S 55.6 0.0 1:28.02 arc_prune 32625 root 20 0 0.0m 0.0m 0.0m R 55.6 0.0 1:27.34 arc_prune 22572 root 20 0 0.0m 0.0m 0.0m S 55.3 0.0 10:02.92 arc_prune 31989 root 20 0 0.0m 0.0m 0.0m R 55.3 0.0 1:28.03 arc_prune 3353 root 20 0 0.0m 0.0m 0.0m R 54.9 0.0 8:58.81 arc_prune 10252 root 20 0 0.0m 0.0m 0.0m R 54.9 0.0 2:36.37 arc_prune 1522 root 20 0 0.0m 0.0m 0.0m S 53.9 0.0 158:42.45 arc_prune 3694 root 20 0 0.0m 0.0m 0.0m R 53.9 0.0 1:20.79 arc_prune 13394 root 20 0 0.0m 0.0m 0.0m R 53.9 0.0 10:35.78 arc_prune 24592 root 20 0 0.0m 0.0m 0.0m R 53.9 0.0 1:54.19 arc_prune 25859 root 20 0 0.0m 0.0m 0.0m S 53.9 0.0 1:51.71 arc_prune 8194 root 20 0 0.0m 0.0m 0.0m S 53.6 0.0 0:54.51 arc_prune 18472 root 20 0 0.0m 0.0m 0.0m R 53.6 0.0 2:08.73 arc_prune 29525 root 20 0 0.0m 0.0m 0.0m R 53.6 0.0 1:35.81 arc_prune 32291 root 20 0 0.0m 0.0m 0.0m S 53.6 0.0 1:28.00 arc_prune 3156 root 20 0 0.0m 0.0m 0.0m R 53.3 0.0 3:17.68 arc_prune 6224 root 20 0 0.0m 0.0m 0.0m S 53.3 0.0 1:11.80 arc_prune 9788 root 20 0 0.0m 0.0m 0.0m S 53.3 0.0 0:46.00 arc_prune 10341 root 20 0 0.0m 0.0m 0.0m R 53.3 0.0 2:36.23 arc_prune 11881 root 20 0 0.0m 0.0m 0.0m S 53.0 0.0 2:31.57 arc_prune 24030 root 20 0 0.0m 0.0m 0.0m R 52.6 0.0 1:55.44 arc_prune === zpool iostat -v 1 output (for a while): capacity operations bandwidth pool alloc free read write read write ------------------------------------------- ----- ----- ----- ----- ----- ----- fst 1.15T 679G 68 0 650K 0 mirror 588G 340G 31 0 331K 0 sdj - - 20 0 179K 0 sdk - - 10 0 152K 0 mirror 588G 340G 36 0 319K 0 sdl - - 17 0 132K 0 sdm - - 18 0 187K 0 ------------------------------------------- ----- ----- ----- ----- ----- ----- srv 1.30T 6.86T 2 35 187K 144K mirror 443G 2.29T 0 0 63.8K 0 ata-HGST_HUS724030ALA640_PN2234P8KTWJYY - - 0 0 63.8K 0 ata-HGST_HUS724030ALA640_PN2234P9G620TW - - 0 0 0 0 ata-HGST_HUS724030ALA640_PN2234P9G66E2U - - 0 0 0 0 mirror 443G 2.29T 0 17 0 71.8K ata-HGST_HUS724030ALA640_PN2234P9G69TKU - - 0 5 0 23.9K ata-HGST_HUS724030ALA640_PN2234P9G69TXU - - 0 5 0 23.9K ata-HGST_HUS724030ALA640_PN2234P9G69U2U - - 0 5 0 23.9K mirror 443G 2.29T 1 17 124K 71.8K ata-HGST_HUS724030ALA640_PN2234P9G6EBUU - - 1 5 124K 23.9K ata-HGST_HUS724030ALA640_PN2234P9G6ESAU - - 0 5 0 23.9K ata-HGST_HUS724030ALA640_PN2234P9G6G70U - - 0 5 0 23.9K logs - - - - - - nvme0n1p1 900K 19.9G 0 0 0 0 cache - - - - - - nvme0n1p2 334G 764G 0 0 0 0 ------------------------------------------- ----- ----- ----- ----- ----- ----- capacity operations bandwidth pool alloc free read write read write ------------------------------------------- ----- ----- ----- ----- ----- ----- fst 1.15T 679G 110 0 1.07M 0 mirror 588G 340G 59 0 634K 0 sdj - - 28 0 303K 0 sdk - - 30 0 331K 0 mirror 588G 340G 50 0 459K 0 sdl - - 28 0 303K 0 sdm - - 21 0 155K 0 ------------------------------------------- ----- ----- ----- ----- ----- ----- srv 1.30T 6.86T 2 229 183K 1.00M mirror 443G 2.29T 2 73 183K 335K ata-HGST_HUS724030ALA640_PN2234P8KTWJYY - - 2 24 183K 112K ata-HGST_HUS724030ALA640_PN2234P9G620TW - - 0 24 0 112K ata-HGST_HUS724030ALA640_PN2234P9G66E2U - - 0 23 0 112K mirror 443G 2.29T 0 77 0 347K ata-HGST_HUS724030ALA640_PN2234P9G69TKU - - 0 25 0 116K ata-HGST_HUS724030ALA640_PN2234P9G69TXU - - 0 25 0 116K ata-HGST_HUS724030ALA640_PN2234P9G69U2U - - 0 25 0 116K mirror 443G 2.29T 0 77 0 347K ata-HGST_HUS724030ALA640_PN2234P9G6EBUU - - 0 25 0 116K ata-HGST_HUS724030ALA640_PN2234P9G6ESAU - - 0 25 0 116K ata-HGST_HUS724030ALA640_PN2234P9G6G70U - - 0 25 0 116K logs - - - - - - nvme0n1p1 900K 19.9G 0 0 0 0 cache - - - - - - nvme0n1p2 334G 764G 0 0 0 0 ------------------------------------------- ----- ----- ----- ----- ----- ----- capacity operations bandwidth pool alloc free read write read write ------------------------------------------- ----- ----- ----- ----- ----- ----- fst 1.15T 679G 29 0 403K 0 mirror 588G 340G 12 0 171K 0 sdj - - 7 0 79.7K 0 sdk - - 4 0 91.7K 0 mirror 588G 340G 16 0 231K 0 sdl - - 6 0 128K 0 sdm - - 9 0 104K 0 ------------------------------------------- ----- ----- ----- ----- ----- ----- srv 1.30T 6.86T 0 66 63.8K 359K mirror 443G 2.29T 0 21 0 120K ata-HGST_HUS724030ALA640_PN2234P8KTWJYY - - 0 6 0 39.9K ata-HGST_HUS724030ALA640_PN2234P9G620TW - - 0 7 0 39.9K ata-HGST_HUS724030ALA640_PN2234P9G66E2U - - 0 6 0 39.9K mirror 443G 2.29T 0 21 0 120K ata-HGST_HUS724030ALA640_PN2234P9G69TKU - - 0 7 0 39.9K ata-HGST_HUS724030ALA640_PN2234P9G69TXU - - 0 6 0 39.9K ata-HGST_HUS724030ALA640_PN2234P9G69U2U - - 0 6 0 39.9K mirror 443G 2.29T 0 22 63.8K 120K ata-HGST_HUS724030ALA640_PN2234P9G6EBUU - - 0 7 63.8K 39.9K ata-HGST_HUS724030ALA640_PN2234P9G6ESAU - - 0 6 0 39.9K ata-HGST_HUS724030ALA640_PN2234P9G6G70U - - 0 7 0 39.9K logs - - - - - - nvme0n1p1 900K 19.9G 0 0 0 0 cache - - - - - - nvme0n1p2 334G 764G 0 0 0 0 ------------------------------------------- ----- ----- ----- ----- ----- ----- capacity operations bandwidth pool alloc free read write read write ------------------------------------------- ----- ----- ----- ----- ----- ----- fst 1.15T 679G 97 0 797K 0 mirror 588G 340G 58 0 474K 0 sdj - - 27 0 263K 0 sdk - - 30 0 211K 0 mirror 588G 340G 38 0 323K 0 sdl - - 23 0 203K 0 sdm - - 14 0 120K 0 ------------------------------------------- ----- ----- ----- ----- ----- ----- srv 1.30T 6.86T 2 176 187K 789K mirror 443G 2.29T 0 58 59.8K 263K ata-HGST_HUS724030ALA640_PN2234P8KTWJYY - - 0 19 59.8K 87.7K ata-HGST_HUS724030ALA640_PN2234P9G620TW - - 0 18 0 87.7K ata-HGST_HUS724030ALA640_PN2234P9G66E2U - - 0 19 0 87.7K mirror 443G 2.29T 0 59 0 263K ata-HGST_HUS724030ALA640_PN2234P9G69TKU - - 0 19 0 87.7K ata-HGST_HUS724030ALA640_PN2234P9G69TXU - - 0 19 0 87.7K ata-HGST_HUS724030ALA640_PN2234P9G69U2U - - 0 19 0 87.7K mirror 443G 2.29T 1 57 128K 263K ata-HGST_HUS724030ALA640_PN2234P9G6EBUU - - 1 18 128K 87.7K ata-HGST_HUS724030ALA640_PN2234P9G6ESAU - - 0 19 0 87.7K ata-HGST_HUS724030ALA640_PN2234P9G6G70U - - 0 18 0 87.7K logs - - - - - - nvme0n1p1 900K 19.9G 0 0 0 0 cache - - - - - - nvme0n1p2 334G 764G 0 0 0 0 ------------------------------------------- ----- ----- ----- ----- ----- ----- capacity operations bandwidth pool alloc free read write read write ------------------------------------------- ----- ----- ----- ----- ----- ----- fst 1.15T 679G 70 0 426K 0 mirror 588G 340G 38 0 263K 0 sdj - - 21 0 128K 0 sdk - - 16 0 135K 0 mirror 588G 340G 31 0 163K 0 sdl - - 10 0 67.7K 0 sdm - - 20 0 95.6K 0 ------------------------------------------- ----- ----- ----- ----- ----- ----- srv 1.30T 6.86T 1 46 116K 2.36M mirror 443G 2.29T 0 0 59.8K 0 ata-HGST_HUS724030ALA640_PN2234P8KTWJYY - - 0 0 59.8K 0 ata-HGST_HUS724030ALA640_PN2234P9G620TW - - 0 0 0 0 ata-HGST_HUS724030ALA640_PN2234P9G66E2U - - 0 0 0 0 mirror 443G 2.29T 0 37 0 2.31M ata-HGST_HUS724030ALA640_PN2234P9G69TKU - - 0 12 0 789K ata-HGST_HUS724030ALA640_PN2234P9G69TXU - - 0 11 0 789K ata-HGST_HUS724030ALA640_PN2234P9G69U2U - - 0 12 0 789K mirror 443G 2.29T 0 8 55.8K 47.8K ata-HGST_HUS724030ALA640_PN2234P9G6EBUU - - 0 2 55.8K 15.9K ata-HGST_HUS724030ALA640_PN2234P9G6ESAU - - 0 2 0 15.9K ata-HGST_HUS724030ALA640_PN2234P9G6G70U - - 0 2 0 15.9K logs - - - - - - nvme0n1p1 900K 19.9G 0 0 0 0 cache - - - - - - nvme0n1p2 334G 764G 0 0 0 0 ------------------------------------------- ----- ----- ----- ----- ----- ----- capacity operations bandwidth pool alloc free read write read write ------------------------------------------- ----- ----- ----- ----- ----- ----- fst 1.15T 679G 108 0 614K 0 mirror 588G 340G 50 0 299K 0 sdj - - 32 0 203K 0 sdk - - 17 0 95.6K 0 mirror 588G 340G 57 0 315K 0 sdl - - 30 0 155K 0 sdm - - 26 0 159K 0 ------------------------------------------- ----- ----- ----- ----- ----- ----- srv 1.30T 6.86T 2 68 191K 311K mirror 443G 2.29T 0 8 0 47.8K ata-HGST_HUS724030ALA640_PN2234P8KTWJYY - - 0 2 0 15.9K ata-HGST_HUS724030ALA640_PN2234P9G620TW - - 0 2 0 15.9K ata-HGST_HUS724030ALA640_PN2234P9G66E2U - - 0 2 0 15.9K mirror 443G 2.29T 0 29 0 132K ata-HGST_HUS724030ALA640_PN2234P9G69TKU - - 0 9 0 43.8K ata-HGST_HUS724030ALA640_PN2234P9G69TXU - - 0 9 0 43.8K ata-HGST_HUS724030ALA640_PN2234P9G69U2U - - 0 9 0 43.8K mirror 443G 2.29T 2 29 191K 132K ata-HGST_HUS724030ALA640_PN2234P9G6EBUU - - 2 9 191K 43.8K ata-HGST_HUS724030ALA640_PN2234P9G6ESAU - - 0 9 0 43.8K ata-HGST_HUS724030ALA640_PN2234P9G6G70U - - 0 9 0 43.8K logs - - - - - - nvme0n1p1 900K 19.9G 0 0 0 0 cache - - - - - - nvme0n1p2 334G 764G 0 0 0 0 ------------------------------------------- ----- ----- ----- ----- ----- ----- capacity operations bandwidth pool alloc free read write read write ------------------------------------------- ----- ----- ----- ----- ----- ----- fst 1.15T 679G 66 0 379K 0 mirror 588G 340G 26 0 144K 0 sdj - - 12 0 63.8K 0 sdk - - 13 0 79.7K 0 mirror 588G 340G 39 0 235K 0 sdl - - 19 0 120K 0 sdm - - 19 0 116K 0 ------------------------------------------- ----- ----- ----- ----- ----- ----- srv 1.30T 6.86T 2 166 183K 754K mirror 443G 2.29T 0 55 0 251K ata-HGST_HUS724030ALA640_PN2234P8KTWJYY - - 0 18 0 83.7K ata-HGST_HUS724030ALA640_PN2234P9G620TW - - 0 17 0 83.7K ata-HGST_HUS724030ALA640_PN2234P9G66E2U - - 0 18 0 83.7K mirror 443G 2.29T 0 54 0 251K ata-HGST_HUS724030ALA640_PN2234P9G69TKU - - 0 17 0 83.7K ata-HGST_HUS724030ALA640_PN2234P9G69TXU - - 0 18 0 83.7K ata-HGST_HUS724030ALA640_PN2234P9G69U2U - - 0 17 0 83.7K mirror 443G 2.29T 2 55 183K 251K ata-HGST_HUS724030ALA640_PN2234P9G6EBUU - - 2 18 183K 83.7K ata-HGST_HUS724030ALA640_PN2234P9G6ESAU - - 0 17 0 83.7K ata-HGST_HUS724030ALA640_PN2234P9G6G70U - - 0 18 0 83.7K logs - - - - - - nvme0n1p1 900K 19.9G 0 0 0 0 cache - - - - - - nvme0n1p2 334G 764G 0 0 0 0 ------------------------------------------- ----- ----- ----- ----- ----- ----- capacity operations bandwidth pool alloc free read write read write ------------------------------------------- ----- ----- ----- ----- ----- ----- fst 1.15T 679G 126 0 698K 0 mirror 588G 340G 64 0 335K 0 sdj - - 37 0 195K 0 sdk - - 26 0 140K 0 mirror 588G 340G 61 0 363K 0 sdl - - 34 0 207K 0 sdm - - 26 0 155K 0 ------------------------------------------- ----- ----- ----- ----- ----- ----- srv 1.30T 6.86T 3 274 239K 1.23M mirror 443G 2.29T 1 91 120K 418K ata-HGST_HUS724030ALA640_PN2234P8KTWJYY - - 1 30 120K 139K ata-HGST_HUS724030ALA640_PN2234P9G620TW - - 0 29 0 139K ata-HGST_HUS724030ALA640_PN2234P9G66E2U - - 0 30 0 139K mirror 443G 2.29T 0 91 0 418K ata-HGST_HUS724030ALA640_PN2234P9G69TKU - - 0 30 0 139K ata-HGST_HUS724030ALA640_PN2234P9G69TXU - - 0 29 0 139K ata-HGST_HUS724030ALA640_PN2234P9G69U2U - - 0 30 0 139K mirror 443G 2.29T 1 91 120K 418K ata-HGST_HUS724030ALA640_PN2234P9G6EBUU - - 1 30 120K 139K ata-HGST_HUS724030ALA640_PN2234P9G6ESAU - - 0 30 0 139K ata-HGST_HUS724030ALA640_PN2234P9G6G70U - - 0 29 0 139K logs - - - - - - nvme0n1p1 900K 19.9G 0 0 0 0 cache - - - - - - nvme0n1p2 334G 764G 0 0 0 0 ------------------------------------------- ----- ----- ----- ----- ----- ----- capacity operations bandwidth pool alloc free read write read write ------------------------------------------- ----- ----- ----- ----- ----- ----- fst 1.15T 679G 70 0 442K 0 mirror 588G 340G 36 0 215K 0 sdj - - 18 0 95.6K 0 sdk - - 17 0 119K 0 mirror 588G 340G 33 0 227K 0 sdl - - 15 0 123K 0 sdm - - 17 0 104K 0 ------------------------------------------- ----- ----- ----- ----- ----- ----- srv 1.30T 6.86T 2 0 187K 0 mirror 443G 2.29T 0 0 63.7K 0 ata-HGST_HUS724030ALA640_PN2234P8KTWJYY - - 0 0 63.7K 0 ata-HGST_HUS724030ALA640_PN2234P9G620TW - - 0 0 0 0 ata-HGST_HUS724030ALA640_PN2234P9G66E2U - - 0 0 0 0 mirror 443G 2.29T 0 0 0 0 ata-HGST_HUS724030ALA640_PN2234P9G69TKU - - 0 0 0 0 ata-HGST_HUS724030ALA640_PN2234P9G69TXU - - 0 0 0 0 ata-HGST_HUS724030ALA640_PN2234P9G69U2U - - 0 0 0 0 mirror 443G 2.29T 1 0 123K 0 ata-HGST_HUS724030ALA640_PN2234P9G6EBUU - - 1 0 123K 0 ata-HGST_HUS724030ALA640_PN2234P9G6ESAU - - 0 0 0 0 ata-HGST_HUS724030ALA640_PN2234P9G6G70U - - 0 0 0 0 logs - - - - - - nvme0n1p1 900K 19.9G 0 0 0 0 cache - - - - - - nvme0n1p2 334G 764G 0 0 0 0 ------------------------------------------- ----- ----- ----- ----- ----- ----- === arc_summary ------------------------------------------------------------------------ ZFS Subsystem Report Wed Feb 06 16:34:09 2019 ARC Summary: (HEALTHY) Memory Throttle Count: 0 ARC Misc: Deleted: 142.80m Mutex Misses: 14.89m Evict Skips: 21.87b ARC Size: 86.88% 54.65 GiB Target Size: (Adaptive) 78.61% 49.45 GiB Min Size (Hard Limit): 6.25% 3.93 GiB Max Size (High Water): 16:1 62.90 GiB ARC Size Breakdown: Recently Used Cache Size: 25.05% 5.55 GiB Frequently Used Cache Size: 74.95% 16.62 GiB ARC Hash Breakdown: Elements Max: 7.46m Elements Current: 60.33% 4.50m Collisions: 43.79m Chain Max: 8 Chains: 504.81k ARC Total accesses: 2.51b Cache Hit Ratio: 95.24% 2.39b Cache Miss Ratio: 4.76% 119.32m Actual Hit Ratio: 92.99% 2.33b Data Demand Efficiency: 94.16% 486.53m Data Prefetch Efficiency: 86.47% 29.68m CACHE HITS BY CACHE LIST: Anonymously Used: 2.20% 52.51m Most Recently Used: 37.35% 891.38m Most Frequently Used: 60.29% 1.44b Most Recently Used Ghost: 0.10% 2.32m Most Frequently Used Ghost: 0.06% 1.45m CACHE HITS BY DATA TYPE: Demand Data: 19.20% 458.13m Prefetch Data: 1.08% 25.66m Demand Metadata: 78.22% 1.87b Prefetch Metadata: 1.51% 36.00m CACHE MISSES BY DATA TYPE: Demand Data: 23.80% 28.40m Prefetch Data: 3.37% 4.02m Demand Metadata: 66.03% 78.79m Prefetch Metadata: 6.80% 8.12m L2 ARC Summary: (HEALTHY) Low Memory Aborts: 233 Free on Write: 27.52k R/W Clashes: 0 Bad Checksums: 0 IO Errors: 0 L2 ARC Size: (Adaptive) 364.94 GiB Compressed: 91.59% 334.23 GiB Header Size: 0.08% 307.98 MiB L2 ARC Breakdown: 119.32m Hit Ratio: 1.42% 1.69m Miss Ratio: 98.58% 117.63m Feeds: 6.01m L2 ARC Writes: Writes Sent: 100.00% 279.55k DMU Prefetch Efficiency: 1.89b Hit Ratio: 2.24% 42.49m Miss Ratio: 97.76% 1.85b ZFS Tunable: dbuf_cache_hiwater_pct 10 dbuf_cache_lowater_pct 10 dbuf_cache_max_bytes 104857600 dbuf_cache_max_shift 5 dmu_object_alloc_chunk_shift 7 ignore_hole_birth 1 l2arc_feed_again 1 l2arc_feed_min_ms 200 l2arc_feed_secs 1 l2arc_headroom 2 l2arc_headroom_boost 200 l2arc_noprefetch 1 l2arc_norw 0 l2arc_write_boost 8388608 l2arc_write_max 8388608 metaslab_aliquot 524288 metaslab_bias_enabled 1 metaslab_debug_load 0 metaslab_debug_unload 0 metaslab_fragmentation_factor_enabled 1 metaslab_lba_weighting_enabled 1 metaslab_preload_enabled 1 metaslabs_per_vdev 200 send_holes_without_birth_time 1 spa_asize_inflation 24 spa_config_path /etc/zfs/zpool.cache spa_load_verify_data 1 spa_load_verify_maxinflight 10000 spa_load_verify_metadata 1 spa_slop_shift 5 zfetch_array_rd_sz 1048576 zfetch_max_distance 8388608 zfetch_max_streams 8 zfetch_min_sec_reap 2 zfs_abd_scatter_enabled 1 zfs_abd_scatter_max_order 10 zfs_admin_snapshot 1 zfs_arc_average_blocksize 8192 zfs_arc_dnode_limit 0 zfs_arc_dnode_limit_percent 10 zfs_arc_dnode_reduce_percent 10 zfs_arc_grow_retry 0 zfs_arc_lotsfree_percent 10 zfs_arc_max 0 zfs_arc_meta_adjust_restarts 4096 zfs_arc_meta_limit 0 zfs_arc_meta_limit_percent 75 zfs_arc_meta_min 0 zfs_arc_meta_prune 10000 zfs_arc_meta_strategy 1 zfs_arc_min 0 zfs_arc_min_prefetch_lifespan 0 zfs_arc_p_aggressive_disable 1 zfs_arc_p_dampener_disable 1 zfs_arc_p_min_shift 0 zfs_arc_pc_percent 0 zfs_arc_shrink_shift 0 zfs_arc_sys_free 0 zfs_autoimport_disable 1 zfs_compressed_arc_enabled 1 zfs_dbgmsg_enable 0 zfs_dbgmsg_maxsize 4194304 zfs_dbuf_state_index 0 zfs_deadman_checktime_ms 5000 zfs_deadman_enabled 1 zfs_deadman_synctime_ms 1000000 zfs_dedup_prefetch 0 zfs_delay_min_dirty_percent 60 zfs_delay_scale 500000 zfs_delete_blocks 20480 zfs_dirty_data_max 4294967296 zfs_dirty_data_max_max 4294967296 zfs_dirty_data_max_max_percent 25 zfs_dirty_data_max_percent 10 zfs_dirty_data_sync 67108864 zfs_dmu_offset_next_sync 0 zfs_expire_snapshot 300 zfs_flags 0 zfs_free_bpobj_enabled 1 zfs_free_leak_on_eio 0 zfs_free_max_blocks 100000 zfs_free_min_time_ms 1000 zfs_immediate_write_sz 32768 zfs_max_recordsize 1048576 zfs_mdcomp_disable 0 zfs_metaslab_fragmentation_threshold 70 zfs_metaslab_segment_weight_enabled 1 zfs_metaslab_switch_threshold 2 zfs_mg_fragmentation_threshold 85 zfs_mg_noalloc_threshold 0 zfs_multihost_fail_intervals 5 zfs_multihost_history 0 zfs_multihost_import_intervals 10 zfs_multihost_interval 1000 zfs_multilist_num_sublists 0 zfs_no_scrub_io 0 zfs_no_scrub_prefetch 0 zfs_nocacheflush 0 zfs_nopwrite_enabled 1 zfs_object_mutex_size 64 zfs_pd_bytes_max 52428800 zfs_per_txg_dirty_frees_percent 30 zfs_prefetch_disable 0 zfs_read_chunk_size 1048576 zfs_read_history 0 zfs_read_history_hits 0 zfs_recover 0 zfs_resilver_delay 2 zfs_resilver_min_time_ms 3000 zfs_scan_idle 50 zfs_scan_min_time_ms 1000 zfs_scrub_delay 4 zfs_send_corrupt_data 0 zfs_sync_pass_deferred_free 2 zfs_sync_pass_dont_compress 5 zfs_sync_pass_rewrite 2 zfs_sync_taskq_batch_pct 75 zfs_top_maxinflight 32 zfs_txg_history 0 zfs_txg_timeout 5 zfs_vdev_aggregation_limit 131072 zfs_vdev_async_read_max_active 3 zfs_vdev_async_read_min_active 1 zfs_vdev_async_write_active_max_dirty_percent 60 zfs_vdev_async_write_active_min_dirty_percent 30 zfs_vdev_async_write_max_active 10 zfs_vdev_async_write_min_active 2 zfs_vdev_cache_bshift 16 zfs_vdev_cache_max 16384 zfs_vdev_cache_size 0 zfs_vdev_max_active 1000 zfs_vdev_mirror_non_rotating_inc 0 zfs_vdev_mirror_non_rotating_seek_inc 1 zfs_vdev_mirror_rotating_inc 0 zfs_vdev_mirror_rotating_seek_inc 5 zfs_vdev_mirror_rotating_seek_offset 1048576 zfs_vdev_queue_depth_pct 1000 zfs_vdev_raidz_impl [fastest] original scalar sse2 ssse3 avx2 zfs_vdev_read_gap_limit 32768 zfs_vdev_scheduler noop zfs_vdev_scrub_max_active 2 zfs_vdev_scrub_min_active 1 zfs_vdev_sync_read_max_active 10 zfs_vdev_sync_read_min_active 10 zfs_vdev_sync_write_max_active 10 zfs_vdev_sync_write_min_active 10 zfs_vdev_write_gap_limit 4096 zfs_zevent_cols 80 zfs_zevent_console 0 zfs_zevent_len_max 512 zfs_zil_clean_taskq_maxalloc 1048576 zfs_zil_clean_taskq_minalloc 1024 zfs_zil_clean_taskq_nthr_pct 100 zil_replay_disable 0 zil_slog_bulk 786432 zio_delay_max 30000 zio_dva_throttle_enabled 1 zio_requeue_io_start_cut_in_line 1 zio_taskq_batch_pct 75 zvol_inhibit_dev 0 zvol_major 230 zvol_max_discard_blocks 16384 zvol_prefetch_bytes 131072 zvol_request_sync 0 zvol_threads 32 zvol_volmode 1 Thanks ProblemType: Bug DistroRelease: Ubuntu 18.04 Package: zfsutils-linux 0.7.5-1ubuntu16.4 ProcVersionSignature: Ubuntu 4.15.0-39.42-generic 4.15.18 Uname: Linux 4.15.0-39-generic x86_64 NonfreeKernelModules: zfs zunicode zavl icp zcommon znvpair ApportVersion: 2.20.9-0ubuntu7.5 Architecture: amd64 Date: Wed Feb 6 16:26:17 2019 InstallationDate: Installed on 2016-04-04 (1038 days ago) InstallationMedia: Ubuntu-Server 16.04 LTS "Xenial Xerus" - Beta amd64 (20160325) ProcEnviron: TERM=rxvt-unicode-256color PATH=(custom, no user) XDG_RUNTIME_DIR=<set> LANG=en_US.UTF-8 SHELL=/bin/bash SourcePackage: zfs-linux UpgradeStatus: Upgraded to bionic on 2018-08-16 (174 days ago) modified.conffile..etc.sudoers.d.zfs: [inaccessible: [Errno 13] Permission denied: '/etc/sudoers.d/zfs'] To manage notifications about this bug go to: https://bugs.launchpad.net/zfs/+bug/1814983/+subscriptions -- Mailing list: https://launchpad.net/~kernel-packages Post to : kernel-packages@lists.launchpad.net Unsubscribe : https://launchpad.net/~kernel-packages More help : https://help.launchpad.net/ListHelp