Hi Alexander,

thanks for you answer.
What you wrote in terms of postgres I knew. I just tested to log all
statements with statistics. This is a lot of unstructured data in a logfile.
But this is the best I found as far.

The database is running on a solaris box. So DTrace is no problem. I
couldn't find any dtrace scripts for postgres. Do you know any scripts
except this sample script?

Thanks.
Uwe

On Sun, Jun 15, 2008 at 4:03 PM, Alexander Staubo <[EMAIL PROTECTED]> wrote:

> On Sun, Jun 15, 2008 at 3:48 PM, Uwe Bartels <[EMAIL PROTECTED]>
> wrote:
> > Tuning a sql statements I'm familiar with. Finding a sql statement which
> > takes too long due to i/o is probably easy as well. But how about
> statements
> > that take about 100 ms, that read a lot and that are executed several
> times
> > per second?
>
> Take a look at the PostgreSQL manual chapter on monitoring and statistics:
>
>  http://www.postgresql.org/docs/8.3/interactive/monitoring.html
>
> If you have access to DTrace (available on Solaris, OS X and possibly
> FreeBSD), you could hook the low-level system calls to reads and
> writes. If you don't have access to DTrace, the pg_statio_* set of
> tables is your main option. In particular, pg_statio_user_tables and
> pg_statio_user_indexes. See the documentation for the meaning of the
> individual columns.
>
> Unfortunately, the statistics tables are not transaction-specific
> (indeed I believe they only update once you commit the transaction,
> and then only after a delay), meaning they capture statistics about
> everything currently going on in the database. The only way to capture
> statistics about a single query, then, is to run it in complete
> isolation.
>
> Alexander.
>

Reply via email to