Thanks for the input. This was not a case of degraded vdev, but only a missing log device (which i cannot get rid of..). I'll try offlining some vdevs and see what happens - altough this should be automatic atf all times IMO. On Jun 30, 2011 1:25 PM, "Markus Kovero" <markus.kov...@nebula.fi> wrote: > > >> To me it seems that writes are not directed properly to the devices that have most free space - almost exactly the opposite. The writes seem to go to the devices that have _least_ free space, instead of the devices that have most free space. The same effect that can be seen in these 60s averages can also be observed in a shorter timespan, like a second or so. > >> Is there something obvious I'm missing? > > > Not sure how OI should behave, I've managed to even writes & space usage between vdevs by bringing device offline in vdev you don't want to writes end up to. > If you have degraded vdev in your pool, zfs will try not to write there, and this may be the case here as well as I don't see zpool status output. > > Yours > Markus Kovero >
_______________________________________________ zfs-discuss mailing list zfs-discuss@opensolaris.org http://mail.opensolaris.org/mailman/listinfo/zfs-discuss