On Sun, Dec 30, 2012 at 06:02:40PM +0100, Eugen Leitl wrote: > > Happy $holidays, > > I have a pool of 8x ST31000340AS on an LSI 8-port adapter as
Just a little update on the home NAS project. I've set the pool sync to disabled, and added a couple of 8. c4t1d0 <ATA-INTELSSDSA2M080-02G9 cyl 11710 alt 2 hd 224 sec 56> /pci@0,0/pci1462,7720@11/disk@1,0 9. c4t2d0 <ATA-INTELSSDSA2M080-02G9 cyl 11710 alt 2 hd 224 sec 56> /pci@0,0/pci1462,7720@11/disk@2,0 I had no clue what the partitions names (created with napp-it web interface, a la 5% log and 95% cache, of 80 GByte) were and so did a iostat -xnp 1.4 0.3 5.5 0.0 0.0 0.0 0.0 0.0 0 0 c4t1d0 0.1 0.0 3.7 0.0 0.0 0.0 0.0 0.5 0 0 c4t1d0s2 0.1 0.0 2.6 0.0 0.0 0.0 0.0 0.5 0 0 c4t1d0s8 0.0 0.0 0.0 0.0 0.0 0.0 0.0 0.2 0 0 c4t1d0p0 0.0 0.0 0.0 0.0 0.0 0.0 0.0 0.0 0 0 c4t1d0p1 0.0 0.0 0.0 0.0 0.0 0.0 0.0 0.0 0 0 c4t1d0p2 0.0 0.0 0.0 0.0 0.0 0.0 0.0 0.0 0 0 c4t1d0p3 0.0 0.0 0.0 0.0 0.0 0.0 0.0 0.0 0 0 c4t1d0p4 1.2 0.3 1.4 0.0 0.0 0.0 0.0 0.0 0 0 c4t2d0 0.0 0.0 0.6 0.0 0.0 0.0 0.0 0.4 0 0 c4t2d0s2 0.0 0.0 0.7 0.0 0.0 0.0 0.0 0.4 0 0 c4t2d0s8 0.1 0.0 0.0 0.0 0.0 0.0 0.0 0.2 0 0 c4t2d0p0 0.0 0.0 0.0 0.0 0.0 0.0 0.0 0.0 0 0 c4t2d0p1 0.0 0.0 0.0 0.0 0.0 0.0 0.0 0.0 0 0 c4t2d0p2 then issued # zpool add tank0 cache /dev/dsk/c4t1d0p1 /dev/dsk/c4t2d0p1 # zpool add tank0 log mirror /dev/dsk/c4t1d0p0 /dev/dsk/c4t2d0p0 which resulted in root@oizfs:~# zpool status pool: rpool state: ONLINE scan: scrub repaired 0 in 0h1m with 0 errors on Wed Jan 2 21:09:23 2013 config: NAME STATE READ WRITE CKSUM rpool ONLINE 0 0 0 c4t3d0s0 ONLINE 0 0 0 errors: No known data errors pool: tank0 state: ONLINE scan: scrub repaired 0 in 5h17m with 0 errors on Wed Jan 2 17:53:20 2013 config: NAME STATE READ WRITE CKSUM tank0 ONLINE 0 0 0 raidz3-0 ONLINE 0 0 0 c3t5000C500098BE9DDd0 ONLINE 0 0 0 c3t5000C50009C72C48d0 ONLINE 0 0 0 c3t5000C50009C73968d0 ONLINE 0 0 0 c3t5000C5000FD2E794d0 ONLINE 0 0 0 c3t5000C5000FD37075d0 ONLINE 0 0 0 c3t5000C5000FD39D53d0 ONLINE 0 0 0 c3t5000C5000FD3BC10d0 ONLINE 0 0 0 c3t5000C5000FD3E8A7d0 ONLINE 0 0 0 logs mirror-1 ONLINE 0 0 0 c4t1d0p0 ONLINE 0 0 0 c4t2d0p0 ONLINE 0 0 0 cache c4t1d0p1 ONLINE 0 0 0 c4t2d0p1 ONLINE 0 0 0 errors: No known data errors which resulted in bonnie++ befo': NAME SIZE Bonnie Date(y.m.d) File Seq-Wr-Chr %CPU Seq-Write %CPU Seq-Rewr %CPU Seq-Rd-Chr %CPU Seq-Read %CPU Rnd Seeks %CPU Files Seq-Create Rnd-Create rpool 59.5G start 2012.12.28 15576M 24 MB/s 61 47 MB/s 18 40 MB/s 19 26 MB/s 98 273 MB/s 48 2657.2/s 25 16 12984/s 12058/s tank0 7.25T start 2012.12.29 15576M 35 MB/s 86 145 MB/s 48 109 MB/s 50 25 MB/s 97 291 MB/s 53 819.9/s 12 16 12634/s 9194/s aftuh: -Wr-Chr %CPU Seq-Write %CPU Seq-Rewr %CPU Seq-Rd-Chr %CPU Seq-Read %CPU Rnd Seeks %CPU Files Seq-Create Rnd-Create rpool 59.5G start 2012.12.28 15576M 24 MB/s 61 47 MB/s 18 40 MB/s 19 26 MB/s 98 273 MB/s 48 2657.2/s 25 16 12984/s 12058/s tank0 7.25T start 2013.01.03 15576M 35 MB/s 86 149 MB/s 48 111 MB/s 50 26 MB/s 98 404 MB/s 76 1094.3/s 12 16 12601/s 9937/s Does the layout make sense? Do the stats make sense, or is there still something very wrong with that pool? Thanks. _______________________________________________ zfs-discuss mailing list zfs-discuss@opensolaris.org http://mail.opensolaris.org/mailman/listinfo/zfs-discuss