> hi, > have installed two identical servers with ubuntu 12.04, and installed > drbd8-utils and ocfs2-tools. > > i have configured drbd.conf as: >
> # You can find an example in /usr/share/doc/drbd.../drbd.conf.example > > include "drbd.d/global_common.conf"; > include "drbd.d/*.res"; > > resource datacluster { > protocol C; > > handlers { > pri-on-incon-degr "echo o > /proc/sysrq-trigger ; halt -f"; > pri-lost-after-sb "echo o > /proc/sysrq-trigger ; halt -f"; > local-io-error "echo o > /proc/sysrq-trigger ; halt -f"; > outdate-peer "/usr/lib/heartbeat/drbd-peer-outdater -t 5"; > } > > startup { > wfc-timeout 120; > degr-wfc-timeout 120; > become-primary-on both; > } > > disk { > on-io-error detach; > fencing resource-and-stonith; > } > > net { > allow-two-primaries; > cram-hmac-alg sha1; > shared-secret "password"; > after-sb-0pri discard-zero-changes; > after-sb-1pri discard-secondary; > after-sb-2pri disconnect; > } > > syncer { > rate 100M; > verify-alg sha1; > al-extents 257; > } > > on nas1 { > device /dev/drbd0; > disk /dev/xvdb2; > address 192.168.10.21:7788; > meta-disk /dev/xvdb1[0]; > } > > on nas2 { > device /dev/drbd0; > disk /dev/xvdb2; > address 192.168.10.22:7788; > meta-disk /dev/xvdb1[0]; > } > } > > > and configured cluster.conf as: > > cluster: > node_count = 2 > name = dcl > > node: > ip_port = 7777 > ip_address = 192.168.10.21 > number = 1 > name = nas1 > cluster = dcl > > node: > ip_port = 7777 > ip_address = 192.168.10.22 > number = 2 > name = nas2 > cluster = dcl > > > > everything appears to be working fine. drbd-overview shows the device is > connected as primary on both servers, i've not noticed any problems at any > time with the drbd configuration. > > the drbd0 device was formatted using mkfs.ocfs2 -T mail /dev/drbd0 > again everything seems fines. > i can mount the device (mount -t ocfs2 /dev/drbd0 /var/clustermount) on > either server with no problems, unmount it, mount it on the other server, > unmount it. i can do this repeatedly without errors. and echo "slotmap" | > debugfs.ocfs2 -n /dev/drdb0 show the slotmap exactly as expected at any > point. > > when i have the cluster mounted on nas1, and try to simultaneously mount > it on nas2, it takes a long time, causes disk activity (always on the same > disk) to go through the roof, then i get a pingack not received in time > message, nas1 is forcibly rebooted, and /dev/drbd0 gets successfully > mounted on nas2. > when nas1 is rebooted, i can mount it, and the slotmap shows the cluster > successfully mounted on both servers. > if i try mounting to both servers in any of way it fails, eg, mount to > nas2 first, then mount to nas1, nas2 gets rebooted, it may or may not > succesfully mount on nas1, if it was successful, trying to mount it on nas2 > after it's reboot will forcibly reboot nas1. > > anyone have any idea what could be going on here? cause it's doing my head > in. > > just in case it's relevant, these 2 servers are virtualized servers on > xenserver 6.2 for testing purposes, once testing is successful the > configuration will be installed on identical physical servers. > > also can anyone confirm, if i get this working properly, if i export the > ocfs2 cluster to an 20 node apache server farm using nfs, concurrent writes > to /var/www/vhosts/*/logs/*.log will work ok? or will each apache node need > to be a part of the cluster configuration? > > > thanks > lee. > >
_______________________________________________ Ocfs2-users mailing list Ocfs2-users@oss.oracle.com https://oss.oracle.com/mailman/listinfo/ocfs2-users