On Wed, Apr 25, 2018 at 2:21 AM, Jonathan Rudenberg
<jonat...@titanous.com> wrote:
> This issue happened again in production, here are the stack traces from three 
> we grabbed before nuking the >400 hanging backends.
>
> [...]
> #4  0x000055fccb93b21c in LWLockAcquire+188() at 
> /usr/lib/postgresql/10/bin/postgres at lwlock.c:1233
> #5  0x000055fccb925fa7 in dsm_create+151() at 
> /usr/lib/postgresql/10/bin/postgres at dsm.c:493
> #6  0x000055fccb6f2a6f in InitializeParallelDSM+511() at 
> /usr/lib/postgresql/10/bin/postgres at parallel.c:266
> [...]

Thank you.  These stacks are all blocked trying to acquire
DynamicSharedMemoryControlLock.  My theory is that they can't because
one backend -- the one that emitted the error "FATAL:  cannot unpin a
segment that is not pinned" -- is deadlocked against itself.  After
emitting that error you can see from Andreas's "seabisquit" stack that
that shmem_exit() runs dsm_backend_shutdown() which runs dsm_detach()
which tries to acquire DynamicSharedMemoryControlLock again, even
though we already hold it at that point.

I'll write a patch to fix that unpleasant symptom.  While holding
DynamicSharedMemoryControlLock we shouldn't raise any errors without
releasing it first, because the error handling path will try to
acquire it again.  That's a horrible failure mode as you have
discovered.

But that isn't the root problem: we shouldn't be raising that error,
and I'd love to see the stack of the one process that did that and
then self-deadlocked.  I will have another go at trying to reproduce
it here today.

-- 
Thomas Munro
http://www.enterprisedb.com

Reply via email to