Hmm. I don't have a lot of ideas, I'm just adding lockdep, splice and FS people (and Oleg, just because) to the cc to see if anybody else does.
Al, Peter? So the problematic op *seems* to be the splice into /proc/<pid>/attr/ files, which causes that "pipe -> cred_guard_mutex" locking. While the *normal* ordering would be the other way around, coming from execve(), which has the cred_guard_mutex -> VFS locks ordering for reading the executable headers. Al, can we break either of those? Do we need to hold on to the cred mutex that long? We get it fairly early (prepare_bprm_creds) and we drop it very late (install_exec_creds), which means that it covers a lot. But that seems pretty basic. The splice into /proc/<pid>/attr/* seems to be the more annoying one, and maybe we just shouldn't allow splicing into or from /proc? Dave, is this new (it doesn't *smell* new to me), or is it just that trinity is doing new splice things? Or is the XFS i_iolock required for this thing to happen at all? Adding Ben Myers to the cc just for luck/completeness. Linus On Mon, Jul 15, 2013 at 6:53 PM, Dave Jones <da...@redhat.com> wrote: > > [ 696.047396] ====================================================== > [ 696.049036] [ INFO: possible circular locking dependency detected ] > [ 696.050689] 3.11.0-rc1+ #53 Tainted: G W > [ 696.052182] ------------------------------------------------------- > [ 696.053846] trinity-child2/14017 is trying to acquire lock: > [ 696.055429] (&sig->cred_guard_mutex){+.+.+.}, at: [<ffffffff8122fa55>] > proc_pid_attr_write+0xf5/0x140 > [ 696.057652] but task is already holding lock: > [ 696.060171] (&pipe->mutex/1){+.+.+.}, at: [<ffffffff811c7096>] > pipe_lock+0x26/0x30 > [ 696.062097] which lock already depends on the new lock. > [ 696.065723] the existing dependency chain (in reverse order) is: > [ 696.068295] -> #2 (&pipe->mutex/1){+.+.+.}: > [ 696.070650] [<ffffffff810c4bd1>] lock_acquire+0x91/0x1f0 > [ 696.072258] [<ffffffff8170677a>] mutex_lock_nested+0x7a/0x410 > [ 696.073916] [<ffffffff811c7096>] pipe_lock+0x26/0x30 > [ 696.075464] [<ffffffff811f0024>] > generic_file_splice_write+0x64/0x170 > [ 696.077192] [<ffffffffa001afc0>] xfs_file_splice_write+0xb0/0x230 > [xfs] > [ 696.078914] [<ffffffff811f1f5a>] SyS_splice+0x24a/0x7e0 > [ 696.080455] [<ffffffff81714694>] tracesys+0xdd/0xe2 > [ 696.082000] -> #1 (&(&ip->i_iolock)->mr_lock){++++++}: > [ 696.084428] [<ffffffff810c4bd1>] lock_acquire+0x91/0x1f0 > [ 696.086017] [<ffffffff8108bc92>] down_read_nested+0x52/0xa0 > [ 696.087643] [<ffffffffa007b310>] xfs_ilock+0x1d0/0x280 [xfs] > [ 696.089305] [<ffffffffa001aa12>] xfs_file_aio_read+0x112/0x3e0 [xfs] > [ 696.091037] [<ffffffff811be2c0>] do_sync_read+0x80/0xb0 > [ 696.092571] [<ffffffff811be931>] vfs_read+0xa1/0x170 > [ 696.094124] [<ffffffff811c4ae1>] kernel_read+0x41/0x60 > [ 696.095693] [<ffffffff811c4d03>] prepare_binprm+0xb3/0x130 > [ 696.097307] [<ffffffff811c6b29>] > do_execve_common.isra.29+0x599/0x6c0 > [ 696.099036] [<ffffffff811c6f36>] SyS_execve+0x36/0x50 > [ 696.100577] [<ffffffff81714a79>] stub_execve+0x69/0xa0 > [ 696.102157] -> #0 (&sig->cred_guard_mutex){+.+.+.}: > [ 696.104536] [<ffffffff810c40b6>] __lock_acquire+0x1786/0x1af0 > [ 696.106189] [<ffffffff810c4bd1>] lock_acquire+0x91/0x1f0 > [ 696.107813] [<ffffffff817062b5>] > mutex_lock_interruptible_nested+0x75/0x4c0 > [ 696.109616] [<ffffffff8122fa55>] proc_pid_attr_write+0xf5/0x140 > [ 696.111243] [<ffffffff811bf372>] __kernel_write+0x72/0x150 > [ 696.112870] [<ffffffff811f03f9>] write_pipe_buf+0x49/0x70 > [ 696.114503] [<ffffffff811efa24>] splice_from_pipe_feed+0x84/0x140 > [ 696.116202] [<ffffffff811efdce>] __splice_from_pipe+0x6e/0x90 > [ 696.117836] [<ffffffff811f18c1>] splice_from_pipe+0x51/0x70 > [ 696.119419] [<ffffffff811f1919>] default_file_splice_write+0x19/0x30 > [ 696.121147] [<ffffffff811f1f5a>] SyS_splice+0x24a/0x7e0 > [ 696.122719] [<ffffffff81714694>] tracesys+0xdd/0xe2 > [ 696.124272] other info that might help us debug this: > > [ 696.127604] Chain exists of: &sig->cred_guard_mutex --> > &(&ip->i_iolock)->mr_lock --> &pipe->mutex/1 > > [ 696.131445] Possible unsafe locking scenario: > > [ 696.133763] CPU0 CPU1 > [ 696.135108] ---- ---- > [ 696.136420] lock(&pipe->mutex/1); > [ 696.137745] lock(&(&ip->i_iolock)->mr_lock); > [ 696.139469] lock(&pipe->mutex/1); > [ 696.141030] lock(&sig->cred_guard_mutex); > [ 696.142409] > *** DEADLOCK *** > > [ 696.145237] 2 locks held by trinity-child2/14017: > [ 696.146532] #0: (sb_writers#4){.+.+.+}, at: [<ffffffff811f2441>] > SyS_splice+0x731/0x7e0 > [ 696.148435] #1: (&pipe->mutex/1){+.+.+.}, at: [<ffffffff811c7096>] > pipe_lock+0x26/0x30 > [ 696.150357] stack backtrace: > [ 696.152287] CPU: 2 PID: 14017 Comm: trinity-child2 Tainted: G W > 3.11.0-rc1+ #53 > [ 696.156695] ffffffff825270a0 ffff8801fee2db90 ffffffff8170164e > ffffffff824f84c0 > [ 696.158392] ffff8801fee2dbd0 ffffffff816fda92 ffff8801fee2dc20 > ffff8802331446d8 > [ 696.160300] ffff880233143fc0 0000000000000002 0000000000000002 > ffff8802331446d8 > [ 696.162217] Call Trace: > [ 696.163292] [<ffffffff8170164e>] dump_stack+0x4e/0x82 > [ 696.164695] [<ffffffff816fda92>] print_circular_bug+0x200/0x20f > [ 696.166194] [<ffffffff810c40b6>] __lock_acquire+0x1786/0x1af0 > [ 696.167677] [<ffffffff810c4bd1>] lock_acquire+0x91/0x1f0 > [ 696.169111] [<ffffffff8122fa55>] ? proc_pid_attr_write+0xf5/0x140 > [ 696.170633] [<ffffffff8122fa55>] ? proc_pid_attr_write+0xf5/0x140 > [ 696.172151] [<ffffffff817062b5>] > mutex_lock_interruptible_nested+0x75/0x4c0 > [ 696.173765] [<ffffffff8122fa55>] ? proc_pid_attr_write+0xf5/0x140 > [ 696.175280] [<ffffffff8122fa55>] ? proc_pid_attr_write+0xf5/0x140 > [ 696.176793] [<ffffffff8119e869>] ? alloc_pages_current+0xa9/0x170 > [ 696.178301] [<ffffffff8122fa55>] proc_pid_attr_write+0xf5/0x140 > [ 696.179799] [<ffffffff811f03b0>] ? splice_direct_to_actor+0x1f0/0x1f0 > [ 696.181346] [<ffffffff811bf372>] __kernel_write+0x72/0x150 > [ 696.182796] [<ffffffff811f03b0>] ? splice_direct_to_actor+0x1f0/0x1f0 > [ 696.184354] [<ffffffff811f03f9>] write_pipe_buf+0x49/0x70 > [ 696.185779] [<ffffffff811efa24>] splice_from_pipe_feed+0x84/0x140 > [ 696.187293] [<ffffffff811f03b0>] ? splice_direct_to_actor+0x1f0/0x1f0 > [ 696.188850] [<ffffffff811efdce>] __splice_from_pipe+0x6e/0x90 > [ 696.190325] [<ffffffff811f03b0>] ? splice_direct_to_actor+0x1f0/0x1f0 > [ 696.191878] [<ffffffff811f18c1>] splice_from_pipe+0x51/0x70 > [ 696.193335] [<ffffffff811f1919>] default_file_splice_write+0x19/0x30 > [ 696.194896] [<ffffffff811f1f5a>] SyS_splice+0x24a/0x7e0 > [ 696.196312] [<ffffffff81714694>] tracesys+0xdd/0xe2 > > -- To unsubscribe from this list: send the line "unsubscribe linux-kernel" in the body of a message to majord...@vger.kernel.org More majordomo info at http://vger.kernel.org/majordomo-info.html Please read the FAQ at http://www.tux.org/lkml/