Hello. I have a geographically distributed cluster, all machines have public IP addresses. No virtual IP subnet exists, so no multicast is available.
I thought that UDPu transport can work in such environment, doesn't it? To test everything in advance, I've set up a corosync+pacemaker on Ubuntu 14.04 with the following corosync.conf: totem { transport: udpu interface { ringnumber: 0 bindnetaddr: ip-address-of-the-current-machine mcastport: 5405 } } nodelist { node { ring0_addr: node1 } node { ring0_addr: node2 } } ... (here node1 and node2 are hostnames from /etc/hosts on both machines). After running "service corosync start; service pacemaker start" logs show no problems, but actually both nodes are always offline: root@node1:/etc/corosync# crm status | grep node OFFLINE: [ node1 node2 ] and "crm node online" (as all other attempts to make crm to do something) are timed out with "communication error". No iptables, selinux, apparmor and other bullshit are active: just pure virtual machines with single public IP addresses on each. Also tcpdump shows that UDB packets on port 5405 are going in and out, and if I e.g. stop corosync at node1, the tcpdump output at node2 changes significantly. So they see each other definitely. And if I attach a gvpe adapter to these 2 machines with a private subnet and switch transport to the default one, corosync + pacemaker begin to work. So my question is: what am I doing wrong? Maybe UDPu is not suitable for communications among machines with public IP addresses only?
_______________________________________________ Pacemaker mailing list: Pacemaker@oss.clusterlabs.org http://oss.clusterlabs.org/mailman/listinfo/pacemaker Project Home: http://www.clusterlabs.org Getting started: http://www.clusterlabs.org/doc/Cluster_from_Scratch.pdf Bugs: http://bugs.clusterlabs.org