http://git.kernel.org/?p=linux/kernel/git/torvalds/linux-2.6.git;a=commit;h=6ca497a83e592d64e050c4d04b6dedb8c915f39a
The race has a tiny window. The backdoor read (via nfs file handle) has to happen after a file has been deleted (and committed to journal) and before the journal has checkpointed (or replayed). I am surprised you are hitting it so often. Are you doing a lot of deletes? Sunil michael.a.jaqu...@verizon.com wrote: > Are there any workarounds besides the nordirplus option on the nfs clients? > > -Mike Jaquays > Office: 972-718-2982 > Cell: 214-587-3882 > > -----Original Message----- > From: Sunil Mushran [mailto:sunil.mush...@oracle.com] > Sent: Thursday, February 18, 2010 11:47 AM > To: Jaquays, Michael A. > Cc: ocfs2-users@oss.oracle.com > Subject: Re: [Ocfs2-users] Kernel Panic ocfs2_inode_lock_full > > Yes, this is a known issue. Only occurs when nfs is in the equation. > This issue has been fixed in mainline quite some time ago. We are > in the process of backporting that to 1.4. > > michael.a.jaqu...@verizon.com wrote: >> All, >> >> I have a 3 node cluster that is experiencing kernel panics once every few >> days. We are sharing some of the ocfs2 filesystems via nfs to some web app >> servers. The app servers mount the filesystems with the nordirplus option. >> Are there any known pitfalls with using an nfs4 server and ocfs2? I haven't >> seen a case where all three nodes are down at the same time, but the issue >> seems to travel from node to node. Here are the node details: >> >> OS: RHEL5.4 >> Kernel: 2.6.18-164.11.1.el5 #1 SMP Wed Jan 6 13:26:04 EST 2010 x86_64 >> x86_64 x86_64 GNU/Linux >> >> OCFS2 Packages: ocfs2console-1.4.3-1.el5 >> ocfs2-tools-1.4.3-1.el5 >> ocfs2-2.6.18-164.11.1.el5-1.4.4-1.el5 >> >> >> The following is always logged in /var/log/messages right before the node >> panics: >> >> kernel: (11915,0):ocfs2_inode_lock_update:1970 ERROR: bug expression: >> inode->i_generation != le32_to_cpu(fe->i_generation) >> >> kernel: (11915,0):ocfs2_inode_lock_update:1970 ERROR: Invalid dinode 446146 >> disk generation: 1276645928 inode->i_generation: 1276645926 >> >> kernel: ----------- [cut here ] --------- [please bite here ] --------- >> >> The following is part of the kernel panic: >> >> Call Trace: >> [<ffffffff885a2940>] :ocfs2:ocfs2_delete_inode+0x187/0x73f >> [<ffffffff885a27b9>] :ocfs2:ocfs2_delete_inode+0x0/0x73f >> [<ffffffff8002f463>] generic_delete_inode+0xc6/0x143 >> [<ffffffff885a22e3>] :ocfs2:ocfs2_drop_inode+0xca/0x12b >> [<ffffffff885a693f>] :ocfs2:ocfs2_complete_recovery+0x77e/0x910 >> [<ffffffff885a61c1>] :ocfs2:ocfs2_complete_recovery+0x0/0x910 >> [<ffffffff8004d8ed>] run_workqueue+0x94/0xe4 >> [<ffffffff8004a12f>] worker_thread+0x0/0x122 >> [<ffffffff8009fe9f>] keventd_create_kthread+0x0/0xc4 >> [<ffffffff8004a21f>] worker_thread+0xf0/0x122 >> [<ffffffff8008c86c>] default_wake_function+0x0/0xe >> [<ffffffff8009fe9f>] keventd_create_kthread+0x0/0xc4 >> [<ffffffff80032950>] kthread+0xfe/0x132 >> [<ffffffff8005dfb1>] child_rip+0xa/0x11 >> [<ffffffff8009fe9f>] keventd_create_kthread+0x0/0xc4 >> [<ffffffff80032852>] kthread+0x0/0x132 >> [<ffffffff8005dfa6>] child_rip+0x0/0x11 >> >> >> >> Code: 0f 0b 68 1b 3d 5c 88 c2 b2 07 48 83 7b 48 00 75 0a f6 43 2c >> RIP [<ffffffff885928f5>] :ocfs2:ocfs2_inode_lock_full+0x99e/0xe3c >> RSP <ffff810c0af0fc70> >> <0>Kernel panic - not syncing: Fatal exception >> >> Any help anyone could provide would be appreciated. >> >> >> -Mike >> >> _______________________________________________ >> Ocfs2-users mailing list >> Ocfs2-users@oss.oracle.com >> http://oss.oracle.com/mailman/listinfo/ocfs2-users >> >> > > _______________________________________________ Ocfs2-users mailing list Ocfs2-users@oss.oracle.com http://oss.oracle.com/mailman/listinfo/ocfs2-users