RE: all the nost are not reacheable when running massive deletes

2016-04-05 Thread Paco Trujillo
Hi Alain - Over use the cluster was one thing which I was thinking about, and I have requested two new nodes (anyway it was something already planned). But the pattern of nodes with high CPU load is only visible in 1 or two of the nodes, the rest are working correctly. That made me th

RE: all the nost are not reacheable when running massive deletes

2016-04-05 Thread Paco Trujillo
Hi daemeon We have check network and it is ok, in fact the nodes are connecting between themselves with a dedicated network. From: daemeon reiydelle [mailto:daeme...@gmail.com] Sent: maandag 4 april 2016 18:42 To: user@cassandra.apache.org Subject: Re: all the nost are not reacheable when runnin

Clustering key and secondary index behavior changed between 2.0.11 and 3.3.0

2016-04-05 Thread julien muller
Hello, I noticed the following change in behavior while migrating from 2.0.11: Elements of the clustering key seems to not be secondary indexable anymore. Anyone could give me an insight about this issue? Or point me to documents relative to this evolution? There is a sample table with some data:

Re: Clustering key and secondary index behavior changed between 2.0.11 and 3.3.0

2016-04-05 Thread Sylvain Lebresne
I'm surprised this would have fall through the cracks but that certainly look like a regression (a bug). If you can reproduce on 3.0.4 (just to make sure we haven't fixed it recently), then please open a ticket in https://issues.apache.org/jira/browse/CASSANDRA/ with your repro steps. On Tue, Apr

Re: all the nost are not reacheable when running massive deletes

2016-04-05 Thread Alain RODRIGUEZ
> > Over use the cluster was one thing which I was thinking about, and I > have requested two new nodes (anyway it was something already planned). But > the pattern of nodes with high CPU load is only visible in 1 or two of the > nodes, the rest are working correctly. That made me think that addin

RE: cross DC data sync starts without rebuilding nodes on new DC

2016-04-05 Thread SEAN_R_DURITY
What do you mean by “automatically starts syncing?” Are you seeing streaming of existing data or just the addition of new, incoming data? Do you have repairs running as part of your automated maintenance, perhaps? I would expect that new, incoming data would be replicated to the new DC after th

Re: cross DC data sync starts without rebuilding nodes on new DC

2016-04-05 Thread Alain RODRIGUEZ
Hi Vijay. After completing step 7.a (which is altering keyspace with desired > replication factor for new DC) data automatically starts syncing from > existing DC to new DC > My guess: what you are seeing is not data syncing. Well it is, but not old data being streamed but new writes being replic

Re : Optimum value for native_transport_max_concurrent_connections and native_transport_max_concurrent_connections_per_ip

2016-04-05 Thread sai krishnam raju potturi
hi; we are upgrading our cluster from apache-cassandra 2.0.14 to 2.0.17. We have been facing SYN flooding issue (port 9042) in our current version of Cassandra at times. We are hoping to tackle the SYN flooding issues with the following attributes in the YAML file for 2.0.17 native_transport_max

Re: Is it possible to achieve "sticky" request routing?

2016-04-05 Thread Jonathan Haddad
Why is this a requirement? Honestly I don't know why you would do this. On Sat, Apr 2, 2016 at 8:06 PM Mukil Kesavan wrote: > Hello, > > We currently have 3 Cassandra servers running in a single datacenter with > a replication factor of 3 for our keyspace. We also use the SimpleSnitch > wiith D

Re: Is it possible to achieve "sticky" request routing?

2016-04-05 Thread David King
Many C* versions back I did this by writing a custom snitch. This was to maximise use of the row cache but had the effect of pinning requests for a key to a given server. It sounds like you want to do the same thing but ignoring the key. In more modern reversions I believe this may be done as a

Re: Is it possible to achieve "sticky" request routing?

2016-04-05 Thread Steve Robenalt
Aside from Jon's "why" question, I would point out that this only really works because you are running a 3 node cluster with RF=3. If your cluster is going to grow, you can't guarantee that any one server would have all records. I'd be pretty hesitant to put an invisible constraint like that on a c

Re: Is it possible to achieve "sticky" request routing?

2016-04-05 Thread Jonathan Haddad
Yep - Steve hit the nail on the head. The odds of hitting the right server with "sticky routing" goes down as your cluster size increases. You end up adding extra network hops instead of using token aware routing. Unless you're trying to do a coordinator tier (and you're not, according to your o

Re: Is it possible to achieve "sticky" request routing?

2016-04-05 Thread Jim Ancona
Jon and Steve: I don't understand your point. The TokenAwareLoadBalancer identifies the nodes in the cluster that own the data for a particular token and route requests to one of them. As I understand it, the OP wants to send requests for a particular token to the same node every time (assuming it

Re: Cassandra table limitation

2016-04-05 Thread Kai Wang
Once a while the question about table count rises in this list. The most recent is https://groups.google.com/forum/#!topic/nosql-databases/IblAhiLUXdk In short C* is not designed to scale with the table count. For one each table/CF has some fixed memory footprint on *ALL* nodes. The consensus is y

Re: Is it possible to achieve "sticky" request routing?

2016-04-05 Thread Jonathan Haddad
Jim, that's not what he asked. He asked for the equivalent of a load balancer with sticky sessions. On Tue, Apr 5, 2016 at 2:24 PM Jim Ancona wrote: > Jon and Steve: > > I don't understand your point. The TokenAwareLoadBalancer identifies the > nodes in the cluster that own the data for a parti

Re: Is it possible to achieve "sticky" request routing?

2016-04-05 Thread Mukil Kesavan
John and Steve, We use QUORUM consistency for both READS and WRITES. So we won't have the problem of having to pick the right server. The real reason we have this requirement is because we run in a sometimes overloaded virtualized environment that causes the servers to have non-trivial time drifts

Re: Is it possible to achieve "sticky" request routing?

2016-04-05 Thread Steve Robenalt
Hi Micky, The only issues I've seen personally with dropped updates due to even small amounts of time skew were ones I was able to fix by the judicious use of BatchStatements. This was particularly true with counter fields in the 2.0 release (i.e. before counters were fixed), but would also happen

Re: Cassandra table limitation

2016-04-05 Thread Jack Krupansky
What is the nature of these tenants? Are they each creating their own data models? Is there one central authority that will approve of all data models and who can adjust the cluster configuration to support those models? Generally speaking, multi-tenancy is an anti-pattern for Cassandra and for mo

Re: Cassandra table limitation

2016-04-05 Thread jason zhao yang
Thanks Kai, One approach discussed in that post is about disabling slab allocation. What are the consequences except for lower GC performance? Kai Wang 于2016年4月6日周三 上午5:40写道: > Once a while the question about table count rises in this list. The most > recent is > https://groups.google.com/forum

Re: Cassandra table limitation

2016-04-05 Thread jason zhao yang
Hi Jack, Thanks for the reply. Each tenant will has around 50-100 tables for their applications. probably log collection, probably account table, it's not fixed and depends on tenants' need. There will be a team in charge of helping tenant to do data modeling and access patterns. Tenants will no

Re: cross DC data sync starts without rebuilding nodes on new DC

2016-04-05 Thread Vijay Patil
Thanks Alain and Sean, your detailed explanation answers my question. Yes, nodetool status reflects new DC and nodetool netstats says not "No Streams". My all writes going to old DC with local_quorum. Yes this new data might be getting synced into new DC (repair was not running anywhere). I will p

Re: Cassandra table limitation

2016-04-05 Thread Kai Wang
With small data size and unknown access pattern, any particular reason to choose C*? It sounds like a relational database fits better. On Tue, Apr 5, 2016 at 11:40 PM, jason zhao yang < zhaoyangsingap...@gmail.com> wrote: > Hi Jack, > > Thanks for the reply. > > Each tenant will has around 50-100

nodetool drain running for days

2016-04-05 Thread Paco Trujillo
We are having performance problems with our cluster regarding to timeouts when repairs are running or massive deletes. One of the advice I received was update our casssandra version from 2.0.17 to 2.2. I am draining one of the nodes to start the upgrade and the drain is running now for two days.