Hi folks,

using drbd umounting /data1 takes >50 seconds, even though the file
system (ext4, noatime, default) wasn't accessed for more than 2h.
umount ran with 100% CPU load.

# sync
# time umount /data1

real    0m52.772s
user    0m0.000s
sys     0m52.740s


This appears to be a pretty long time. I am concerned that there
is some data sleeping in a buffer that gets flushed only at umount
time.

Kernel is version 4.18.0-0.bpo.1-amd64 on Debian Stretch. drbdutils
is 8.9.10-2. drbd.conf is attached. The bond2 interface used for
drbd synchronization is based upon 2 * 10 Gbit/sec NICs.


Every insightful comment is highly appreciated.

Regards
Harri
#
# see http://www.drbd.org/users-guide-8.4/re-drbdconf.html
# https://www.linbit.com/en/drbd-sync-rate-controller-2/
#
# /etc/hosts
# ~~~~~~~~~~~
# br0 ("external" network interface)
# 192.168.96.184        srvl060a.example.com srvl060a
# 192.168.96.185        srvl060b.example.com srvl060b
#
# bond2 ("internal" network interface, used for drbd synchronization)
# 10.0.0.2      srvl060a.internal
# 10.0.0.3      srvl060b.internal
#
common {
        disk {
                # on-io-error   detach;                 # continue in diskless 
mode (default)
                fencing resource-only;                  # use fence-peer 
handler 
                resync-rate     512M;                   # synchronization rate 
over a dedicated line (4 each disk!)
                c-plan-ahead    0;
                al-extents 1237;                        # activity log extents
        }

        net {
                protocol C;
                max-buffers     32k;
                max-epoch-size  20000;
                sndbuf-size     256k;
                rcvbuf-size     512k;
                # allow-two-primaries;
                cram-hmac-alg sha256;
                # generated using "openssl rand -base64 30":
                shared-secret "YTbwx3QGiBf2xDrWN+9VppvsrOTXFKRy67x3FAu0";
                after-sb-0pri disconnect;
                after-sb-1pri disconnect;
                after-sb-2pri disconnect;
                rr-conflict disconnect;
        }

        startup {
                wfc-timeout 300;
                degr-wfc-timeout 30;
                outdated-wfc-timeout 30;
        }

        handlers {
                pri-on-incon-degr "echo o > /proc/sysrq-trigger ; halt -f";
                pri-lost-after-sb "echo o > /proc/sysrq-trigger ; halt -f";
                fence-peer "/usr/lib/drbd/outdate-peer.sh on 
srvl060a.example.com 192.168.96.185 10.0.0.3 on srvl060b.example.com 
192.168.96.184 10.0.0.2";
                local-io-error "echo o > /proc/sysrq-trigger ; halt -f";

                initial-split-brain "/usr/lib/drbd/notify-split-brain.sh";
                split-brain "/usr/lib/drbd/notify-split-brain.sh";
        }
}

resource data1 {
        device          /dev/drbd1;
        disk            /dev/disk/by-partlabel/data1;
        meta-disk       internal;
        on srvl060a.example.com {
                address         10.0.0.2:7788;
        }
        on srvl060b.example.com {
                address         10.0.0.3:7788;
        }
        startup {
                become-primary-on srvl060a.example.com;
        }
}

resource data2 {
        device          /dev/drbd2;
        disk            /dev/disk/by-partlabel/data2;
        meta-disk       internal;
        on srvl060a.example.com {
                address         10.0.0.2:7789;
        }
        on srvl060b.example.com {
                address         10.0.0.3:7789;
        }
        startup {
                become-primary-on srvl060b.example.com;
        }
}
_______________________________________________
drbd-user mailing list
[email protected]
http://lists.linbit.com/mailman/listinfo/drbd-user

Reply via email to