Throwing my $ .002 Rpc_address defaults to localhosts that if not set, picks value from dns/hostname file. May be you can try setting rpc_address: <to correct pod ip> - see if this helps!
From: Marek Kadek -T (mkadek - CONSOL PARTNERS LTD at Cisco) [mailto:[email protected]] Sent: Wednesday, December 06, 2017 2:19 AM To: [email protected] Subject: Re: Connection refused - 127.0.0.1-Gossip Thanks for any ideas/hints, any straw is worth checking at this point ☺ Well, the clusters “work”, data is correctly stored and queries. I’m interested in why it tries to open a gossip to localhost, and what kind of (performance) impact could this have on clusters. The env vars are correctly passed, and cassandra yaml seems to be correctly set. We are using Cassandra docker image. listen_address: 100.110.253.6 (correct pod ip) # listen_interface: eth0 # listen_interface_prefer_ipv6: false broadcast_rpc_address: 100.110.253.6 It’s also observable with minikube and single C* node on local machine. From: Lerh Chuan Low <[email protected]<mailto:[email protected]>> Reply-To: "[email protected]<mailto:[email protected]>" <[email protected]<mailto:[email protected]>> Date: Tuesday, December 5, 2017 at 11:14 PM To: "[email protected]<mailto:[email protected]>" <[email protected]<mailto:[email protected]>> Subject: Re: Connection refused - 127.0.0.1-Gossip I think as Jeff mentioned it sounds like a configuration issue, are you sure you are using the same configmap/however it's being passed in and just throwing out ideas, maybe the pods are behind a http proxy and you may have forgotten to pass in the env vars? On 6 December 2017 at 08:45, Jeff Jirsa <[email protected]<mailto:[email protected]>> wrote: I don't have any k8 clusters to test with, but do you know how your yaml translates to cassandra.yaml ? What are the listen/broadcast addresses being set? On Tue, Dec 5, 2017 at 6:09 AM, Marek Kadek -T (mkadek - CONSOL PARTNERS LTD at Cisco) <[email protected]<mailto:[email protected]>> wrote: We are experiencing following issues with Cassandra on our kubernetes clusters: ``` @ kubectl exec -it cassandra-cassandra-0 -- tail /var/log/cassandra/debug.log DEBUG [MessagingService-Outgoing-localhost/127.0.0.1-Gossip] 2017-12-05 09:02:06,560 OutboundTcpConnection.java:545 - Unable to connect to localhost/127.0.0.1<http://127.0.0.1> java.net.ConnectException: Connection refused at sun.nio.ch.Net.connect0(Native Method) ~[na:1.8.0_131] at sun.nio.ch.Net.connect(Net.java:454) ~[na:1.8.0_131] at sun.nio.ch.Net.connect(Net.java:446) ~[na:1.8.0_131] at sun.nio.ch.SocketChannelImpl.connect(SocketChannelImpl.java:648) ~[na:1.8.0_131] at org.apache.cassandra.net<https://urldefense.proofpoint.com/v2/url?u=http-3A__org.apache.cassandra.net&d=DwMGaQ&c=LFYZ-o9_HUMeMTSQicvjIg&r=FsmDztdsVuIKml8IDhdHdg&m=O8QyTZVJIRUNcvKDLpX2IcSY35f_-yEoGJt0ELBa2xE&s=Lqf0BleAj9UkULmHnPY_AdTK0vXmP--1Yh1oSWSbjz0&e=>.OutboundTcpConnectionPool.newSocket(OutboundTcpConnectionPool.java:146) ~[apache-cassandra-3.11.0.jar:3.11.0] at org.apache.cassandra.net<https://urldefense.proofpoint.com/v2/url?u=http-3A__org.apache.cassandra.net&d=DwMGaQ&c=LFYZ-o9_HUMeMTSQicvjIg&r=FsmDztdsVuIKml8IDhdHdg&m=O8QyTZVJIRUNcvKDLpX2IcSY35f_-yEoGJt0ELBa2xE&s=Lqf0BleAj9UkULmHnPY_AdTK0vXmP--1Yh1oSWSbjz0&e=>.OutboundTcpConnectionPool.newSocket(OutboundTcpConnectionPool.java:132) ~[apache-cassandra-3.11.0.jar:3.11.0] at org.apache.cassandra.net<https://urldefense.proofpoint.com/v2/url?u=http-3A__org.apache.cassandra.net&d=DwMGaQ&c=LFYZ-o9_HUMeMTSQicvjIg&r=FsmDztdsVuIKml8IDhdHdg&m=O8QyTZVJIRUNcvKDLpX2IcSY35f_-yEoGJt0ELBa2xE&s=Lqf0BleAj9UkULmHnPY_AdTK0vXmP--1Yh1oSWSbjz0&e=>.OutboundTcpConnection.connect(OutboundTcpConnection.java:433) [apache-cassandra-3.11.0.jar:3.11.0] at org.apache.cassandra.net<https://urldefense.proofpoint.com/v2/url?u=http-3A__org.apache.cassandra.net&d=DwMGaQ&c=LFYZ-o9_HUMeMTSQicvjIg&r=FsmDztdsVuIKml8IDhdHdg&m=O8QyTZVJIRUNcvKDLpX2IcSY35f_-yEoGJt0ELBa2xE&s=Lqf0BleAj9UkULmHnPY_AdTK0vXmP--1Yh1oSWSbjz0&e=>.OutboundTcpConnection.run(OutboundTcpConnection.java:262) [apache-cassandra-3.11.0.jar:3.11.0] ``` Basically, it’s tons and tons of the same message over and over (on all clusters, all C* nodes). It tries roughly 4-5 times a second to open a tcp connection to localhost (?) for gossiping. What we know: - does not happen on Cassandra 3.0.15, but happen on 3.11.1 (same configuration). - does happen even on minikube-single-Cassandra “cluster”. - does not happen on docker-compose Cassandra cluster, only on kubernetes one. Our configuration is pretty much this helm chart: https://github.com/kubernetes/charts/blob/master/incubator/cassandra/values.yaml<https://urldefense.proofpoint.com/v2/url?u=https-3A__github.com_kubernetes_charts_blob_master_incubator_cassandra_values.yaml&d=DwMGaQ&c=LFYZ-o9_HUMeMTSQicvjIg&r=FsmDztdsVuIKml8IDhdHdg&m=O8QyTZVJIRUNcvKDLpX2IcSY35f_-yEoGJt0ELBa2xE&s=MKBAUHsdcYKCpNMrNmDc7Ft_oHXcDHTLdRigdHGJuLM&e=> Do you have any idea what it could be related to?
