On 2012-03-31 08:45 , Zhu Han wrote:
Did you hit the bug here?
https://issues.apache.org/jira/browse/CASSANDRA-4054
Yes looks like it. But what confuses me most is not the sstable2json bug
but why the major compaction does not replace the deleted row data with
a tombstone.
Is that a bug or
Hi,
We are in the early stages of thinking about a project that needs to store
data that will be accessed by Hadoop. One of the concerns we have is around
the Latency of HDFS as our use case is is not for reading all the data and
hence we will need custom RecordReaders etc.
I've seen a couple of
Hi,
I've been using cassandra for a while, but after a upgrade to 1.0.7,
every machine kept running perfectly. Well, except one that constantly
throws this error:
ERROR [ReplicateOnWriteStage:39] 2012-04-02 12:02:55,131
AbstractCassandraDaemon.java (line 139) Fatal exception in thread
Threa
Hi,
Bit of a silly question, is row iteration using the RandomPartitioner
deterministic? I don't particularly care what the order is relative to
the row keys (obviously there isn't one, it's the RandomPartitioner),
but if I run a full iteration over all rows in a CF twice, assuming no
underlying
Correct. Random partitioner order is md5 token order. If you make no changes
you will get the same order
On Apr 2, 2012, at 7:53 AM, wrote:
> Hi,
>
> Bit of a silly question, is row iteration using the RandomPartitioner
> deterministic? I don't particularly care what the order is relative
Hi all
Is there a way I can easily retrieve a Merkle tree for a CF, like the one
created during a repair?
I didn't see anything about this in the Thrift API docs, I'm assuming this
is a data structure made available only to internal Cassandra functions.
I would like to explore using the Merkle tr
For a thrift client, you need the following jars at a minimum
apache-cassandra-clientutil-*.jar
apache-cassandra-thrift-*.jar
libthrift-*.jar
slf4j-api-*.jar
slf4j-log4j12-*.jar
all of these jars can be found in the cassandra distribution.
On 04/02/2012 07:40 AM, Rishabh Agrawal wrote:
Any
I didn't fine slf4j files in distribution. So I downloaded them can you help me
how to configure it.
From: Dave Brosius [mailto:dbros...@mebigfatguy.com]
Sent: Monday, April 02, 2012 6:28 PM
To: user@cassandra.apache.org
Subject: Re: Using Thrift
For a thrift client, you need the following jars
Best to read about maven. Save you some grief.
On Apr 2, 2012 3:05 PM, "Rishabh Agrawal"
wrote:
> I didn’t fine slf4j files in distribution. So I downloaded them can you
> help me how to configure it.
>
>
>
> *From:* Dave Brosius [mailto:dbros...@mebigfatguy.com]
> *Sent:* Monday, April 02, 201
slf4j is just a logging facade, if you actually want log files, you need
a logger, say log4j-*.jar in your classpath. Then just configure that
with a log4j.properties file. That properties file also needs to be on
the classpath.
On 04/02/2012 09:05 AM, Rishabh Agrawal wrote:
I didn't fine
???
2012/4/1 juan quintero
>
--
Everton Lima Aleixo
Bacharel em Ciencia da Computação
Universidade Federal de Goiás
There's an open issue for that:
https://issues.apache.org/jira/browse/CASSANDRA-3676
Patch welcome :)
--
Sylvain
On Sat, Mar 31, 2012 at 8:55 PM, Ben McCann wrote:
> I'm trying to upgrade Solandra to use 1.1.0-beta2 and think I found a minor
> issue:
> java.lang.NoClassDefFoundError: edu/stanfor
No. Each agent is responsible for collecting and reporting all the
statistics for the node it is installed on, so there shouldn't be any
duplication.
On Sat, Mar 31, 2012 at 5:01 AM, R. Verlangen wrote:
> Nick, would that also result in useless duplicates of the statistics?
>
>
> 2012/3/30 Nick B
Hi,
I was curious if I compress my data on the client side with Snappy whether
there's any difference between doing that and doing it on the server side?
The wiki said that compression works best where each row has the same
columns. Does this mean the compression will be more efficient on the
se
This is all very hypothetical, but I've been bitten by this before.
Does row_loaded happen to be a binary or boolean value? If so the
secondary index generated by Cassandra will have at most 2 rows, and
they'll be REALLY wide if you have a lot of entries. Since Cassandra
doesn't distribute colum
The server side compression can compress across columns/rows so it will most
likely be more efficient.
Whether you are CPU bound or IO bound depends on your application and node
setup. Unless your working set fits in memory you will be IO bound, and in
that case server side compression helps be
Hi,
how do you select between client- and serverside compression? i'm using
hector and i set compression when creating a cf, so the compression
executes when inserting the data "on the server" oO
greetings, martin
Am 02.04.2012 17:42, schrieb Ben McCann:
Hi,
I was curious if I compress my
Thanks Jeremiah, that's what I has suspected. I appreciate the
confirmation.
Martin, there's not built-in support for doing compression client side, but
it'd be easy for me to do manually since I just have one column with all my
serialized data, which is why I was considering it.
On Mon, Apr 2,
Hi there,
I have a question about the nodetool snapshot.
Situation:
- 3 node cluster
- RF = 3
- fully consistent (not measured, but let's say it is)
Is it true that when I take a snaphot at only one of the 3 nodes this
contains all the data in the cluster (at least 1 replica)?
With kind regards
Is it possible to fetch a column based on the row key and the column’s
timestamp only (not using the column’s name)?
--
View this message in context:
http://cassandra-user-incubator-apache-org.3065146.n2.nabble.com/column-s-timestamp-tp7429905p7429905.html
Sent from the cassandra-u...@incubator.a
On Mon, Apr 2, 2012 at 11:24 AM, Avi-h wrote:
> Is it possible to fetch a column based on the row key and the column’s
> timestamp only (not using the column’s name)?
>
No, but most clients support including the timestamp in the result set, so
you can filter the columns by timestamp from there.
Hi,
What about using a ts as column name and do a get sliced instead ?
--Original Message--
From: Avi-h
To: cassandra-u...@incubator.apache.org
ReplyTo: user@cassandra.apache.org
Subject: column’s timestamp
Sent: Apr 2, 2012 18:24
Is it possible to fetch a column based on the row key an
On Mon, Mar 26, 2012 at 3:31 PM, Deno Vichas > but what if i already
have a bunch (8g per node) data that i need and i
> don't have a way to re-create it.
Note that the below may have unintended consequences if using Counter
column families. It actually can be done with the cluster running,
below
On Mon, Apr 2, 2012 at 9:19 AM, R. Verlangen wrote:
> - 3 node cluster
> - RF = 3
> - fully consistent (not measured, but let's say it is)
>
> Is it true that when I take a snaphot at only one of the 3 nodes this
> contains all the data in the cluster (at least 1 replica)?
Yes.
=Rob
--
=Robert
This is a difficult question to answer for a variety of reasons, but I'll
give it a try, maybe it will be helpful, maybe not.
The most obvious problem with this is that Thrift is buffer based, not
streaming. That means that whatever the size of your chunk it needs to
be received, deserialized, and
Can you send us the stack trace which you can find in the hs_err_pid*.log?
is the system memory all used up (free)? any errors in the logs just before
the crash?
Regards,
On Mon, Mar 26, 2012 at 12:35 AM, Maciej Miklas
wrote:
> I have row cache - it's about 20GB big in this case.
> The proble
Yup Jeremiah, I learned a hard lesson on how cassandra behaves when it runs
out of disk space :-S.I didn't try the compression, but when it ran out
of disk space, or near running out, compaction would fail because it needs
space to create some tmp data files.
I shall get a tatoo that says keep
Is JNA.jar in the path ?
Cheers
-
Aaron Morton
Freelance Developer
@aaronmorton
http://www.thelastpickle.com
On 2/04/2012, at 10:11 PM, Carlos Juzarte Rolo wrote:
> Hi,
>
> I've been using cassandra for a while, but after a upgrade to 1.0.7, every
> machine kept running perfe
I would recommend starting with a higher level client like Hector or Astyanax
http://wiki.apache.org/cassandra/ClientOptions
They have *a lot* of features and will make it easier to focus on learning how
to use Cassandra. Then when you know what you like or do not like about the
existing client
No it's internal only.
Take a look at o.a.c.service.AntiEntropyService
Cheers
-
Aaron Morton
Freelance Developer
@aaronmorton
http://www.thelastpickle.com
On 3/04/2012, at 12:21 AM, Thomas van Neerijnen wrote:
> Hi all
>
> Is there a way I can easily retrieve a Merkle tree fo
On Tue, Apr 3, 2012 at 4:18 AM, Ben Coverston wrote:
> This is a difficult question to answer for a variety of reasons, but I'll
> give it a try, maybe it will be helpful, maybe not.
>
> The most obvious problem with this is that Thrift is buffer based, not
> streaming. That means that whatever th
I faced the same issue:
You can find the similar issue here.
http://stackoverflow.com/questions/8370365/debugging-bizarre-spring-slf4j-jar-issue
Also, Spring community is acknowledging on the SLF4J Issue here
(commons-logging issue):
http://static.springsource.org/spring/docs/3.0.x/spring-framewo
Cool. Thanks. That should be easy enough to fix :-)
On Mon, Apr 2, 2012 at 8:05 AM, Sylvain Lebresne wrote:
> There's an open issue for that:
> https://issues.apache.org/jira/browse/CASSANDRA-3676
> Patch welcome :)
>
> --
> Sylvain
>
> On Sat, Mar 31, 2012 at 8:55 PM, Ben McCann wrote:
> > I
Hi guys,
We are calculating key cache size right now. There is this column family
with ~ 100 million columns and right now we have the cache size set at 2
million.
I suspect that the active data we got is not all fitting in the 2 million
cache size and we at times are getting query execution time
If you have a workload with overwrites you will end up with some data needing
compaction. Running a nightly manual compaction would remove this, but it will
also soak up some IO so it may not be the best solution.
I do not know if Leveled compaction would result in a smaller disk load for the
Take a look at the key cache hit rate in nodetool cfstats.
One approach is to increase the cache size until you do not see a matching
increase in the hit rate.
> Is there a limit to key cache size? I know that is all taken from heap but
> how much max we can go with setting the key cache size
On Tue, Apr 3, 2012 at 11:49 AM, aaron morton wrote:
> Take a look at the key cache hit rate in nodetool cfstats.
>
> One approach is to increase the cache size until you do not see a matching
> increase in the hit rate.
>
Thanks Aaron, what do you think will be the ideal cache hit ratio where w
Ok, thank you.
2012/4/2 Rob Coli
> On Mon, Apr 2, 2012 at 9:19 AM, R. Verlangen wrote:
> > - 3 node cluster
> > - RF = 3
> > - fully consistent (not measured, but let's say it is)
> >
> > Is it true that when I take a snaphot at only one of the 3 nodes this
> > contains all the data in the clus
38 matches
Mail list logo