Roch Bourbonnais - Performance Engineering wrote:
Tao Chen writes:
 > On 5/12/06, Roch Bourbonnais - Performance Engineering
 > <[EMAIL PROTECTED]> wrote:
 > >
 > >   From: Gregory Shaw <[EMAIL PROTECTED]>
 > >   Regarding directio and quickio, is there a way with ZFS to skip the
 > >   system buffer cache?  I've seen big benefits for using directio when
 > >   the data files have been segregated from the log files.
 > >
 > >
 > > Were the benefits coming from extra concurrency (no single writter lock)
> > Does DIO bypass "writter lock" on Solaris?

Yep.

 > Not on AIX, which uses CIO (concurrent I/O) to bypass managing locks
 > at filesystem level:
 > 
http://oracle.ittoolbox.com/white-papers/improving-database-performance-with-aix-concurrent-io-2582
> > > or avoiding the extra copy to page cache > > Certainly. Also to avoid VM overhead (DB does like raw devices).

OK, but again, is it to avoid badly configured readahead, or get extra concurrency, or something else ? I have a hard time that managing the page cache represents a cost when you compare this to a 5ms I/O.
I think you're missing one other thing - handling the memory overload of having orders of magnitude more accessed data than you have memory. Think about how you can handle having a couple hundred GB of dirty data being written by many threads (say either tablespace creates or temp table creation for a large table join) - fsflush and writebehind et. al. just can't keep up with it. Of course, I know ZFS is "better" but to be useable in those situations it needs to be probably an order of magnitude better or so, and I haven't seen any data on a decently big rig with a proper storage config that shows that it is. I'm not saying it's not, I'm just saying I haven't seen the data. :) Like you said, Roch, I've been down this road before and don't want to go down it again. ;)

- Pete
_______________________________________________
zfs-discuss mailing list
zfs-discuss@opensolaris.org
http://mail.opensolaris.org/mailman/listinfo/zfs-discuss

Reply via email to