Public bug reported: Hi,
Following setup: Two nodes are clustered over openvpn (tap-devices). I have configured a drbd/clvm device /dev/fgs/lv_mail, which was formatted with gfs2. The kernel produces tons of the following: ... [26593.231661] drbd0: local disk flush failed with status -5 ... Yesterday, the gfs-cluster was totally dead and I had to rebuild the complete setup from scratch. I know that gfs2 is not ready for productive usage, but I hoped it would be nearly stable enough for first experiences :-) Here my config: drbd: global { usage-count no; } common { protocol C; handlers { # pri-on-incon-degr "halt -f"; # pri-lost-after-sb "halt -f"; } syncer { rate 125M; } startup { become-primary-on both; wfc-timeout 20; degr-wfc-timeout 10; } disk { } net { allow-two-primaries; after-sb-0pri discard-zero-changes; after-sb-1pri discard-secondary; after-sb-2pri disconnect; } } resource r0 { on srv1 { device /dev/drbd0; disk /dev/sda9; address 192.168.254.1:7780; meta-disk internal; } on srv2 { device /dev/drbd0; disk /dev/sda9; address 192.168.254.2:7780; meta-disk internal; } } Here the cluster.conf: <?xml version="1.0" ?> <cluster alias="gfscluster" config_version="1" name="gfscluster"> <cman expected_votes="1" two_node="1"/> <clusternodes> <clusternode name="node1" nodeid="1" votes="1"> <fence> <method name="human"> <device name="human" ipaddr="192.168.254.1"/> </method> </fence> </clusternode> <clusternode name="node2" nodeid="2" votes="1"> <fence> <method name="human"> <device name="human" ipaddr="192.168.254.2"/> </method> </fence> </clusternode> </clusternodes> <fencedevices> <fencedevice agent="fence_manual" name="human"/> </fencedevices> <rm> <failoverdomains/> <resources/> </rm> <fence_daemon post_fail_delay="0" post_join_delay="3"/> </cluster> cman status: cman_tool status Version: 6.1.0 Config Version: 1 Cluster Name: gfscluster Cluster Id: 41396 Cluster Member: Yes Cluster Generation: 20 Membership state: Cluster-Member Nodes: 2 Expected votes: 1 Total votes: 2 Quorum: 1 Active subsystems: 8 Flags: 2node Dirty Ports Bound: 0 11 Node name: node1 Node ID: 1 Multicast addresses: 239.192.161.86 Node addresses: 192.168.254.1 uname -a: Linux srv1 2.6.24-16-xen #2 SMP Sun Apr 20 05:49:43 JST 2008 i686 GNU/Linux It is the fixed version: linux-image-2.6.24-16-xen_2.6.24-16.30zng1_i386.deb from somewhere here in the bugzilla. The tunnel: tap0 Link encap:Ethernet HWaddr 00:ff:33:7c:52:f8 inet addr:192.168.254.1 Bcast:192.168.254.3 Mask:255.255.255.252 inet6 addr: fe80::2ff:33ff:fe7c:52f8/64 Scope:Link UP BROADCAST RUNNING MULTICAST MTU:1500 Metric:1 RX packets:860591 errors:0 dropped:0 overruns:0 frame:0 TX packets:818507 errors:0 dropped:240 overruns:0 carrier:0 collisions:0 txqueuelen:100 RX bytes:631362573 (602.1 MB) TX bytes:293035854 (279.4 MB) tap0 Link encap:Ethernet HWaddr 00:ff:5f:73:33:dc inet addr:192.168.254.2 Bcast:192.168.254.3 Mask:255.255.255.252 inet6 addr: fe80::2ff:5fff:fe73:33dc/64 Scope:Link UP BROADCAST RUNNING MULTICAST MTU:1500 Metric:1 RX packets:817240 errors:0 dropped:0 overruns:0 frame:0 TX packets:863705 errors:0 dropped:0 overruns:0 carrier:0 collisions:0 txqueuelen:100 RX bytes:292331168 (278.7 MB) TX bytes:634969711 (605.5 MB) It connects two servers inside Hetzner backbone (Germany). Same city, same computer center. Most of the time really stable and fast. So I thought, I give it a try. Please let me know, if more information for this problem is required ** Affects: drbd (Ubuntu) Importance: Undecided Status: New -- [hardy] drbd gfs2 clvm https://bugs.launchpad.net/bugs/231248 You received this bug notification because you are a member of Ubuntu Bugs, which is subscribed to Ubuntu. -- ubuntu-bugs mailing list ubuntu-bugs@lists.ubuntu.com https://lists.ubuntu.com/mailman/listinfo/ubuntu-bugs