Public bug reported:

Hi,

Following setup:

Two nodes are clustered over openvpn (tap-devices). I have configured a
drbd/clvm device /dev/fgs/lv_mail, which was formatted with gfs2. The
kernel produces tons of the following:

...
[26593.231661] drbd0: local disk flush failed with status -5
...

Yesterday, the gfs-cluster was totally dead and I had to rebuild the
complete setup from scratch. I know that gfs2 is not ready for
productive usage, but I hoped it would be nearly stable enough for first
experiences :-)

Here my config:

drbd:

global {
  usage-count no;
}

common {
  protocol C;

  handlers {
    # pri-on-incon-degr "halt -f";
    # pri-lost-after-sb "halt -f";
  }

  syncer {
    rate 125M;
  }

  startup {
    become-primary-on both;
    wfc-timeout       20;
    degr-wfc-timeout  10;
  }

  disk {
  }

  net {
    allow-two-primaries;
    after-sb-0pri discard-zero-changes;
    after-sb-1pri discard-secondary;
    after-sb-2pri disconnect;
  }
}

resource r0 {
  on srv1 {
    device     /dev/drbd0;
    disk       /dev/sda9;
    address    192.168.254.1:7780;
    meta-disk  internal;
  }

  on srv2 {
    device     /dev/drbd0;
    disk       /dev/sda9;
    address    192.168.254.2:7780;
    meta-disk  internal;
  }
}

Here the cluster.conf:

<?xml version="1.0" ?>
<cluster alias="gfscluster" config_version="1" name="gfscluster">
        <cman expected_votes="1" two_node="1"/>
        <clusternodes>
                <clusternode name="node1" nodeid="1" votes="1">
                        <fence>
                                <method name="human">
                                        <device name="human" 
ipaddr="192.168.254.1"/>
                                </method>
                        </fence>
                </clusternode>
                <clusternode name="node2" nodeid="2" votes="1">
                        <fence>
                                <method name="human">
                                        <device name="human" 
ipaddr="192.168.254.2"/>
                                </method>
                        </fence>
                </clusternode>
        </clusternodes>
        <fencedevices>
                <fencedevice agent="fence_manual" name="human"/>
        </fencedevices>
        <rm>
                <failoverdomains/>
                <resources/>
        </rm>
        <fence_daemon post_fail_delay="0" post_join_delay="3"/>
</cluster>

cman status:

cman_tool status
Version: 6.1.0
Config Version: 1
Cluster Name: gfscluster
Cluster Id: 41396
Cluster Member: Yes
Cluster Generation: 20
Membership state: Cluster-Member
Nodes: 2
Expected votes: 1
Total votes: 2
Quorum: 1  
Active subsystems: 8
Flags: 2node Dirty 
Ports Bound: 0 11  
Node name: node1
Node ID: 1
Multicast addresses: 239.192.161.86 
Node addresses: 192.168.254.1 

uname -a:

Linux srv1 2.6.24-16-xen #2 SMP Sun Apr 20 05:49:43 JST 2008 i686
GNU/Linux

It is the fixed version: linux-image-2.6.24-16-xen_2.6.24-16.30zng1_i386.deb
from somewhere here in the bugzilla.

The tunnel:

tap0      Link encap:Ethernet  HWaddr 00:ff:33:7c:52:f8  
          inet addr:192.168.254.1  Bcast:192.168.254.3  Mask:255.255.255.252
          inet6 addr: fe80::2ff:33ff:fe7c:52f8/64 Scope:Link
          UP BROADCAST RUNNING MULTICAST  MTU:1500  Metric:1
          RX packets:860591 errors:0 dropped:0 overruns:0 frame:0
          TX packets:818507 errors:0 dropped:240 overruns:0 carrier:0
          collisions:0 txqueuelen:100 
          RX bytes:631362573 (602.1 MB)  TX bytes:293035854 (279.4 MB)

tap0      Link encap:Ethernet  HWaddr 00:ff:5f:73:33:dc  
          inet addr:192.168.254.2  Bcast:192.168.254.3  Mask:255.255.255.252
          inet6 addr: fe80::2ff:5fff:fe73:33dc/64 Scope:Link
          UP BROADCAST RUNNING MULTICAST  MTU:1500  Metric:1
          RX packets:817240 errors:0 dropped:0 overruns:0 frame:0
          TX packets:863705 errors:0 dropped:0 overruns:0 carrier:0
          collisions:0 txqueuelen:100 
          RX bytes:292331168 (278.7 MB)  TX bytes:634969711 (605.5 MB)

It connects two servers inside Hetzner backbone (Germany). Same city,
same computer center. Most of the time really stable and fast. So I
thought, I give it a try.

Please let me know, if more information for this problem is required

** Affects: drbd (Ubuntu)
     Importance: Undecided
         Status: New

-- 
[hardy] drbd gfs2 clvm
https://bugs.launchpad.net/bugs/231248
You received this bug notification because you are a member of Ubuntu
Bugs, which is subscribed to Ubuntu.

-- 
ubuntu-bugs mailing list
ubuntu-bugs@lists.ubuntu.com
https://lists.ubuntu.com/mailman/listinfo/ubuntu-bugs

Reply via email to