On Sat, Apr 17, 2021 at 05:22:59PM +0200, Pavel Stehule wrote: > so 17. 4. 2021 v 17:09 odesÃlatel Justin Pryzby <pry...@telsasoft.com> napsal: > > > On Sat, Apr 17, 2021 at 04:36:52PM +0200, Pavel Stehule wrote: > > > today I worked on postgres's server used for critical service. Because the > > > application is very specific, we had to do final tuning on production > > > server. I fix lot of queries, but I am not able to detect fast queries > > > that > > > does full scan of middle size tables - to 1M rows. Surely I wouldn't log > > > all queries. Now, there are these queries with freq 10 per sec. > > > > > > Can be nice to have a possibility to set a log of queries that do full > > > scan and read more tuples than is specified limit or that does full scan > > > of > > > specified tables. > > > > > > What do you think about the proposed feature? > > > > Are you able to use auto_explain with auto_explain.log_min_duration ? > > Unfortunately, I cannot use it. This server executes 5K queries per > seconds, and I am afraid to decrease log_min_duration. > > The logs are forwarded to the network and last time, when users played with > it, then they had problems with the network. .. > The fullscan of this table needs about 30ms and has 200K rows. So > decreasing log_min_duration to this value is very risky.
auto_explain.sample_rate should allow setting a sufficiently low value of log_min_duration. It exists since v9.6. -- Justin