You should refer to this
https://engineering.eventbrite.com/what-version-of-cassandra-should-i-run/
Thanks
Rahul
On Thu, Jan 9, 2014 at 8:06 AM, Sanjeeth Kumar wrote:
> Hi all,
> What is the latest stable version of cassandra you have in production ?
> We are migrating a large chunk of our
On 09.01.2014, at 03:36, Sanjeeth Kumar wrote:
> Hi all,
> What is the latest stable version of cassandra you have in production ? We
> are migrating a large chunk of our mysql database to cassandra. I see a lot
> of discussions regarding 1.* versions, but I have not seen / could not find
I would like to +1 Jan.
We are using C* 2.0 and have just gone into production directly supporting
the latest revision of www.nytimes.com.
I avoid new features unless I really need them; we are prepared to read
code and make fixes ourselves if necessary, but it has not been.
Best regards,
Micha
Hi all,
I have a use case with huge data which i am not able to design in cassandra.
Table name : MetricResult
Sample Data :
Metric=Sales, Time=Month, Period=Jan-10, Tag=U.S.A, Tag=Pen, Value=10
Metric=Sales, Time=Month, Period=Jan-10, Tag=U.S.A, Tag=Pencil, Value=20
Metric=Sales, Time=Mo
Hi all,
I want to use sstableloader in order to load 1.0.9 data to a 2.0.* cluster.
I know that the sstable format is incompatible between the two versions.
What are my options?
Is there a tool to upgrade sstables directly without any real nodes
involvement?
--
Or Sher
This sort of work sounds much more like a Hadoop/Hive/Pig type of analysis.
What are your latency requirements on queries? Are they ad-hoc or part of an
application? What is the case where you would need to change an existing value?
If it is write once, then Hadoop/Hive is great, if it changes
I seem to be having a similar issue to
https://www.mail-archive.com/user@cassandra.apache.org/msg33340.html
Basically I have a new node of version 1.2.13, when attempting to
connect to the ring it fails to handshake with other nodes in the cluster.
This is attempting to connect to a 1.1.12 n
Hi all:
We have a Cassandra 1.2.10 pre-production ring integrated with Hadoop and
Pig and we are thinking to upgrade the system from 1.2.10 to 2.0.X ( to
2.0.4 for example) before we pass to production.
What are the Pros and Cons According to your experience?
Many thanks and Regards
So, what are you getting from 2.0 if you aren't using the new
features? Why not stick with 1.2.x?
cheers,
Bruce
On Thu, Jan 9, 2014 at 12:37 PM, Laing, Michael
wrote:
> I would like to +1 Jan.
>
> We are using C* 2.0 and have just gone into production directly supporting
> the latest revision of
On Thu, Jan 9, 2014 at 8:46 AM, Miguel Angel Martin junquera <
mianmarjun.mailingl...@gmail.com> wrote:
> We have a Cassandra 1.2.10 pre-production ring integrated with Hadoop
> and Pig and we are thinking to upgrade the system from 1.2.10 to 2.0.X (
> to 2.0.4 for example) before we pass to
@thunder thanks for guidance queries will be fired by application on this
table when users login and browse the application and also through mobile
apps through webservice. Response needs to be quick as user will be doing
analysis over this data on the fly. Writes also needs to be fast as there
is
On Thu, Jan 9, 2014 at 8:33 AM, Elliot Sumner wrote:
>
> I seem to be having a similar issue to
> https://www.mail-archive.com/user@cassandra.apache.org/msg33340.html
>
> Basically I have a new node of version 1.2.13, when attempting to connect
> to the ring it fails to handshake with other nodes
@thunder It will be write once 80% of time but there can be cases client
makes correction in data and then we need to overwrite that..
Thanks
Naresh
On Thu, Jan 9, 2014 at 11:49 PM, Naresh Yadav wrote:
> @thunder thanks for guidance queries will be fired by application on this
> table when
On Thu, Jan 9, 2014 at 6:54 AM, Or Sher wrote:
> I want to use sstableloader in order to load 1.0.9 data to a 2.0.* cluster.
> I know that the sstable format is incompatible between the two versions.
> What are my options?
>
http://www.palominodb.com/blog/2012/09/25/bulk-loading-options-cassandr
To my eye that looks something what the traditional analytics systems do.
You can check out e.g. Acunu Analytics which uses Cassandra as a backend.
Cheers,
Hannu
2014/1/9 Naresh Yadav
> Hi all,
>
> I have a use case with huge data which i am not able to design in
> cassandra.
>
> Table name :
unsubscribe
Well I think you have essentially time-series data, which C* should handle
well, however I think your "Tag" column is going to cause troubles. C* does
have collection columns, but they are not indexable nor usable in WHERE
clause. Your example has both the uniqueness of the data (primary key) and
q
just send that email to user-unsubscribe@cassandra.apache.orgif still confused
check here http://hadonejob.com/img/full/12598654.jpg - Original Message
-From: "Earl Ruby" >;er...@webcdr.com
Hi All,
Does anyone have any suggestions about how to improve performance in the
below use case?
I have a very simple table with a single Partition Key, and one Cluster
key. My app is periodically writing new entries in the table and deleting
old ones.
There are a lot more reads than writes on t
I don't have specific experience upgrading from 1.x to 2.x but I do have to
say that if you have Pig/Hadoop integration working, go with that. I ran
into many small issues getting the integration working with just the right
version of Hadoop/Pig/and Cassandra 2.0.x
On Jan 9, 2014, at 8:46 AM, Mig
Good: doesn't OOM on smallish machines, can use defaults for almost all
params w good results.
Bad: watch the list like a hawk to avoid problems others have, be aware of
bug fixes, workarounds, and Jira issues.
ml
On Thu, Jan 9, 2014 at 12:58 PM, Bruce Durling wrote:
> So, what are you gettin
Am I correct in understanding that it needs to be run on each node in the
cluster? For example, if I have a three node cluster, I'd have to run:
nodetool -h node-1 flush
nodetool -h node-2 flush
nodetool -h node-3 flush
?
Also, does block until it's done?
My use case is recreating a keyspace.
Michael,
On Thu, Jan 9, 2014 at 10:15 PM, Laing, Michael
wrote:
> Good: doesn't OOM on smallish machines, can use defaults for almost all
> params w good results.
>
> Bad: watch the list like a hawk to avoid problems others have, be aware of
> bug fixes, workarounds, and Jira issues.
That's usef
On Thu, Jan 9, 2014 at 1:42 PM, Charlie Mason wrote:
> There are a lot more reads than writes on this particular table. All of
> the queries are just for the partition key. Most of the queries are for
> partition keys that don't exists, more than 99% of the queries.
>
Reads for partition keys th
+1 to what Ed said.
CFS is a good facilitator for running MR jobs on Cassandra to fill the HDFS
requirement (you just want to run MR, but you don't want the whole Hadoop
stack). The source data for your MR jobs should be in Cassandra KS/CFs.
On Mon, Nov 18, 2013 at 3:21 PM, Edward Capriolo wrote
in 1.2 bloom filters and compression meta data were moved off heap. in 2.0
index samples were also taken off heap.
Row cache has been off heap for a while, and the key cache is still on heap.
Any other components you were interested in ?
The best reference will be the NEWS.txt file and the
Owns is how much of the entire, cluster wide, data set the node has. In both
your examples every node has a full copy of the data.
If you have 6 nodes and RF 3 they would have 50%.
Cheers
-
Aaron Morton
New Zealand
@aaronmorton
Co-Founder & Principal Consultant
Apache Cassand
When Xms and Xmx are the same like this the JVM allocates all the memory, and
then on Linux cassandra will ask the OS to lock that memory so it cannot be
paged out. On windows it’s probably getting paged out.
If you only have 4GB on the box, you probably do not want to run cassandra with
4GB. T
> The spikes in latency don’t seem to be correlated to an increase in reads.
> The cluster’s workload is usually handling a maximum workload of 4200
> reads/sec per node, with writes being significantly less, at ~200/sec per
> node. Usually it will be fine with this, with read latencies at aroun
> Is there some other mechanism for forcing expired data to be removed without
> also compacting? (major compaction having obvious problematic side effects,
> and user defined compaction being significant work to script up).
Tombstone compactions may help here
https://issues.apache.org/jira/brow
On Thu, Jan 9, 2014 at 5:02 PM, Aaron Morton wrote:
> Row cache has been off heap for a while, and the key cache is still on
> heap.
>
Usage of the "off heap" row cache has enough associated heap consumption
that I feel compelled to mention that the name "off heap" is slightly
misleading. Certain
We avoid mixing versions for a long time, but we always upgrade one node and
check the application is happy before proceeding. e.g. wait for 30 minutes
before upgrading the others.
If you snapshot before upgrading, and have to roll back after 30 minutes you
can roll back to the snapshot and us
On 10 Jan 2014, at 3:46 am, Miguel Angel Martin junquera
wrote:
> We have a Cassandra 1.2.10 pre-production ring integrated with Hadoop and
> Pig and we are thinking to upgrade the system from 1.2.10 to 2.0.X ( to
> 2.0.4 for example) before we pass to production.
>
> What are the Pros a
Hello everyone,
I've tried to google all I could and also asking at ServerFault first but
as I got no answer I decided to come to the list.
I have 6 machines that I want to use to make a cluster using Cassandra 2.0
Cassandra start in the machines, however it dies after a while with the
error
ja
@Thunder thanks for suggesting design but my main problem is
indexing/quering dynamic Tag on each row that is main context of each row
and most of queries will include that..
As an alternative to cassandra, i tried Apache Blur, in blur table i am
able to store exact same data and all queries also
35 matches
Mail list logo