Am 08.03.2011 12:48, schrieb Jeremy Chadwick: > On Tue, Mar 08, 2011 at 12:26:49PM +0100, Patrick M. Hausen wrote: >> we use a big JBOD and ZFS with raidz2 as the target >> for our nightly Amanda backups. >> >> I already suspected that the fact that the FS was > 90% full might >> be the cause of our backup performance continously decreasing. >> >> I just added another vdev - 6 disks of 750 GB each, raidz2 and the >> FS usage is back to 71% currently. This was while backups were >> running and write performance instantly skyrocketed compared to >> the values before. >> >> So, is it possible to name a reasonable amount of free space to >> keep on a raidz2 volume? On last year's EuroBSDCon I got >> the impression that with recent (RELENG_8) ZFS merges >> I could get away with using around 90%. > > I'm in no way attempting to dissuade you from your efforts to figure out > a good number for utilisation, but when I hear of disks -- no matter how > many -- being 90% full, I immediately conclude performance is going to > suck simply because the outer "tracks" on a disk contains more sectors > than the inner "tracks". This is the reason for performance degradation > as the seek offset increases, resulting in graphs like this:
Whatever. I've experienced similar massive performance decrease even on a non-redundant single-disk ZFS setup after the ZFS (8-STABLE between 8.0 and before 8.2) had filled up once. Even clearing the disk down to 70% didn't get my /usr (which was a ZFS mount) snappy again. The speed decrease was one to two orders of magnitude in excess of what you'd attribute to the CLV or sectors-per-track change across the disk. What I heard from my 7200/min WD RE3 drive (which seeks rather fast for a 7200/min drive - I think it was the fastest seeking 7200/min drive when I bought it) it was seeking and thrashing heads like hell even on single-threaded bulk reads of large files, and I suppose there was fragmentation and/or non-caching of metadata afoot, and it was far worse than any decrease in constant linear velocity or sectors-per-track of the disk tracks could explain, and the relevant ZFS ARC related options didn't rectify that either, so I reverted to GJOURNAL-enabled UFS which gave me a much better performance on a 5400/min disk than I've ever had with a halfway filled ZFS on the 7200/min RAID-class disk. And bulk transfer rates of both drives are beyond any doubt. In other words, the file system didn't recover speed (I'm not sure if that's a zfs or zpool feature), and I attribute that (and the failure to rm files from a 100% full file system) to the write-ahead-logging behaviour of ZFS. Any comments? -- Matthias Andree _______________________________________________ freebsd-stable@freebsd.org mailing list http://lists.freebsd.org/mailman/listinfo/freebsd-stable To unsubscribe, send any mail to "freebsd-stable-unsubscr...@freebsd.org"