>>> Charles Taylor <[email protected]> schrieb am 17.07.2014 um 17:24 in Nachricht <[email protected]>: > I feel like this is something that must have been covered extensively already > but I've done a lot of googling, looked at a lot of cluster configs, but have > not found the solution. > > I have an HA NFS cluster (corosync+pacemaker). The relevant rpms are listed > below but I'm not sure they are that important to the question which is > this... > > When performing managed failovers of the NFS-exported file system resource > from one node to the other (crm resource move), any active NFS clients > experience an I/O error when the file system is unexported. In other words, > you must unexport it to unmount it. As soon as it is unexported, clients are > no longer able to write to it and experience an I/O error (rather than just > blocking).
Do you hard-mount or soft-mount NFS? Do you use NFSv3 or NFSv4? > > In a failure scenario this is not a problem becuase the file system is never > unexported on the primary server. Rather the server just goes down, the > secondary takes over the resources and client I/O blocks until the process is > complete and then goes about its business. We would like this same behavior > for a *managed* failover but have not found a mount or export option/scenario > that works. Is it possible? What am I missing? > > I realize this is more of an nfs/exportfs question but I would think that > those implementing NFS HA clusters would be familiar with the scenario I'm > describing. > > Regards, > > Charlie Taylor > > pacemaker-cluster-libs-1.1.7-6.el6.x86_64 > pacemaker-cli-1.1.7-6.el6.x86_64 > pacemaker-1.1.7-6.el6.x86_64 > pacemaker-libs-1.1.7-6.el6.x86_64 > resource-agents-3.9.2-40.el6.x86_64 > fence-agents-3.1.5-35.el6.x86_64 > > Red Hat Enterprise Linux Server release 6.3 (Santiago) > > Linux biostor3.ufhpc 2.6.32-279.19.1.el6.x86_64 #1 SMP Sat Nov 24 14:35:28 > EST 2012 x86_64 x86_64 x86_64 GNU/Linux > > [root@biostor4 bs34]# crm status > ============ > Last updated: Thu Jul 17 10:55:04 2014 > Last change: Thu Jul 17 07:59:47 2014 via crmd on biostor3.ufhpc > Stack: openais > Current DC: biostor3.ufhpc - partition with quorum > Version: 1.1.7-6.el6-148fccfd5985c5590cc601123c6c16e966b85d14 > 2 Nodes configured, 2 expected votes > 20 Resources configured. > ============ > > Online: [ biostor3.ufhpc biostor4.ufhpc ] > > Resource Group: grp_b3v0 > vg_b3v0 (ocf::heartbeat:LVM): Started biostor3.ufhpc > fs_b3v0 (ocf::heartbeat:Filesystem): Started biostor3.ufhpc > ip_vbio3 (ocf::heartbeat:IPaddr2): Started biostor3.ufhpc > ex_b3v0_1 (ocf::heartbeat:exportfs): Started biostor3.ufhpc > ex_b3v0_2 (ocf::heartbeat:exportfs): Started biostor3.ufhpc > ex_b3v0_3 (ocf::heartbeat:exportfs): Started biostor3.ufhpc > ex_b3v0_4 (ocf::heartbeat:exportfs): Started biostor3.ufhpc > ex_b3v0_5 (ocf::heartbeat:exportfs): Started biostor3.ufhpc > Resource Group: grp_b4v0 > vg_b4v0 (ocf::heartbeat:LVM): Started biostor4.ufhpc > fs_b4v0 (ocf::heartbeat:Filesystem): Started biostor4.ufhpc > ip_vbio4 (ocf::heartbeat:IPaddr2): Started biostor4.ufhpc > ex_b4v0_1 (ocf::heartbeat:exportfs): Started biostor4.ufhpc > ex_b4v0_2 (ocf::heartbeat:exportfs): Started biostor4.ufhpc > ex_b4v0_3 (ocf::heartbeat:exportfs): Started biostor4.ufhpc > ex_b4v0_4 (ocf::heartbeat:exportfs): Started biostor4.ufhpc > ex_b4v0_5 (ocf::heartbeat:exportfs): Started biostor4.ufhpc > st_bio3 (stonith:fence_ipmilan): Started biostor4.ufhpc > st_bio4 (stonith:fence_ipmilan): Started biostor3.ufhpc > > > > _______________________________________________ > Linux-HA mailing list > [email protected] > http://lists.linux-ha.org/mailman/listinfo/linux-ha > See also: http://linux-ha.org/ReportingProblems _______________________________________________ Linux-HA mailing list [email protected] http://lists.linux-ha.org/mailman/listinfo/linux-ha See also: http://linux-ha.org/ReportingProblems
