čt 27. 2. 2020 v 6:16 odesílatel Kyotaro Horiguchi <horikyota....@gmail.com>
napsal:

> Hello.
>
> At Thu, 27 Feb 2020 10:18:16 +0900, Yugo NAGATA <nag...@sraoss.co.jp>
> wrote in
> > On Wed, 26 Feb 2020 18:51:21 +0100
> > Julien Rouhaud <rjuju...@gmail.com> wrote:
> >
> > > On Thu, Feb 27, 2020 at 02:35:18AM +0900, Yugo NAGATA wrote:
> > > > Hi,
> > > >
> > > > Attached is a patch for allowing auto_explain to log plans before
> > > > queries are executed.
> > > >
> > > > Currently, auto_explain logs plans only after query executions,
> > > > so if a query gets stuck its plan could not be logged. If we can
> > > > know plans of stuck queries, we may get some hints to resolve the
> > > > stuck. This is useful when you are testing and debugging your
> > > > application whose queries get stuck in some situations.
> > >
> > > Indeed that could be useful.
> >
> > > > This patch adds  new option log_before_query to auto_explain.
> > >
> > > Maybe "log_before_execution" would be better?
> >
> > Thanks!  This seems better also to me.
> >
> > >
> > > > Setting auto_explain.log_before_query option logs all plans before
> > > > queries are executed regardless of auto_explain.log_min_duration
> > > > unless this is set -1 to disable logging.  If log_before_query is
> > > > enabled, only duration time is logged after query execution as in
> > > > the case of when both log_statement and log_min_duration_statement
> > > > are enabled.
> > >
> > > I'm not sure about this behavior.  The final explain plan is needed at
> least if
> > > log_analyze, log_buffers or log_timing are enabled.
> >
> > In the current patch, log_before_query (will be log_before_execution)
> > has no effect if log_analyze is enabled in order to avoid to log the
> > same plans twice.  Instead, is it better to log the plan always twice,
> > before and after the execution, if  log_before_query is enabled
> > regardless of log_min_duration or log_analyze?
>
> Honestly, I don't think showing plans for all queries is useful
> behavior.
>
> If you allow the stuck query to be canceled, showing plan in
> PG_FINALLY() block in explain_ExecutorRun would work, which look like
> this.
>
> explain_ExecutorRun()
> {
>   ...
>   PG_TRY();
>   {
>       ...
>       else
>          starndard_ExecutorRun();
>       nesting_level--;
>   }
>   PG_CATCH();
>   {
>       nesting_level--;
>
>       if (auto_explain_log_failed_plan &&
>        <maybe the time elapsed from start exceeds min_duration>)
>       {
>           'show the plan'
>       }
>    }
> }
>
> regards.
>

It can work - but still it is not good enough solution. We need "query
debugger" that allows to get some query execution metrics online.

There was a problem with memory management for passing plans between
processes. Can we used temp files instead shared memory?

Regards

Pavel


> --
> Kyotaro Horiguchi
> NTT Open Source Software Center
>
>
>

Reply via email to