El sáb, 20-08-2011 a las 01:22 +0200, Peter Schuller escribió:
> > Is there any chance that the entire file from source node got streamed to
> > destination node even though only small amount of data in hte file from
> > source node is supposed to be streamed destination node?
>
> Yes, but the thi
Are there any plans to backport this to 0.8?
El mar, 19-07-2011 a las 11:43 -0500, Jonathan Ellis escribió:
> https://issues.apache.org/jira/browse/CASSANDRA-2280
>
> 2011/7/19 Héctor Izquierdo Seliva :
> > Hi all,
> >
> > Maybe I'm doing something wrong, b
Hi all,
Maybe I'm doing something wrong, but calling ./nodetool -h repair
mykeyspace mycolumnfamily should only repair mycolumnfamily right?
Everytime I try a repair it repairs the whole key space instead of just
one column family. I'm on cassandra 0.8.1
> Of course. I wasn't thinking clearly.
>
> So, back to a previous point you brought up, I will have heavy reads and
> even heavier writes. How would you rate the benefits of flashcache in
> such a scenario? Is it still an overall performance boost worth the
> expense?
We have also heavy r
>
> If using the version that has both rt and wt caches, is it just the wt
> cache that's polluted for compactions/flushes? If not, why does the rt
> cache also get polluted?
>
As I said, all reads go through flashcache, so if you read three 10 GB
sstables for a compaction you will get those
> Interesting. So, there is no segregation between read and write cache
> space? A compaction or flush can evict blocks in the read cache if it
> needs the space for write buffering?
There are two versions, the -wt (write through) that will cache also
what is written, and the normal version t
>
> Hector, some before/after numbers would be great if you can find them.
> Thanks!
>
I'll try and get some for you :)
> What happens when your cache gets trashed? Do compactions and flushes
> go slower?
>
If you use flashcache-wt flushed and compacted sstables will go to the
cache.
A
I've been using flashcache for a while in production. It improves read
performance and latency was halved by a good chunk, though I don't
remember the exact numbers.
Problems: compactions will trash your cache, and so will memtable
flushes. Right now there's no way to avoid that.
If you want, I
At the end I had to restart the whole cluster. This is the second time
I've had to do this. Would it be possible to add a command that forces
all nodes to remove all the ring data and start it fresh? I'd rather
have a few seconds of errors in the clients that the two to five minutes
that takes a fu
I'm also having problems with removetoken. Maybe I'm doing it wrong, but
I was under the impression that I just had to call once removetoken.
When I take a look at the nodes ring, the dead node keeps popping up.
What's even more incredible is that in some of them it says UP
epair. Try running it on a single CF to start with.
>
>
> Good luck
>
> -
> Aaron Morton
> Freelance Cassandra Developer
> @aaronmorton
> http://www.thelastpickle.com
>
> On 9 Jul 2011, at 16:45, Héctor Izquierdo Seliva wrote:
>
> > Hi Peter.
> >
>
Hi Peter.
I have a problem with repair, and it's that it always brings the node
doing the repairs down. I've tried setting index_interval to 5000, and
it still dies with OutOfMemory errors, or even worse, it generates
thousands of tiny sstables before dying.
I've tried like 20 repairs during thi
ndra Developer
> @aaronmorton
> http://www.thelastpickle.com
>
> On 8 Jul 2011, at 09:38, Héctor Izquierdo Seliva wrote:
>
> > Hi everyone,
> >
> > I'm having thousands of these errors:
> >
> > WARN [CompactionExecutor:1] 2011-07-08 16:36:45,705
>
Hi everyone,
I'm having thousands of these errors:
WARN [CompactionExecutor:1] 2011-07-08 16:36:45,705
CompactionManager.java (line 737) Non-fatal error reading row
(stacktrace follows)
java.io.IOError: java.io.IOException: Impossible row size
6292724931198053
at
org.apache.cassandra.db.
Forcing a full gc doesn't help either. Now the node is stuck in an
endless loop of full gcs that don't free any memory.
Hi all,
I don't seem to be able to complete a full repair on one of the nodes.
Memory consuptiom keeps growing till it starts complaining about not
having enough heap. I had to disable the automatic memtable flush, as it
was generating thousands of almost empty memtables.
My guess is that the key
> That would be a bug. Can you open a ticket with the exact version
> you're using and
> the circumstance where this happens.
>
> Thanks.
>
> --
> Sylvain
https://issues.apache.org/jira/browse/CASSANDRA-2863
Worker.run(ThreadPoolExecutor.java:908)
at java.lang.Thread.run(Thread.java:662)
El mié, 06-07-2011 a las 11:22 +0200, Sylvain Lebresne escribió:
> 2011/7/6 Héctor Izquierdo Seliva :
> > Hi, i've been struggling to repair my failed node for the past few days,
> >
Hi, i've been struggling to repair my failed node for the past few days,
and I've seen this erros a few times.
java.lang.RuntimeException: Cannot recover SSTable with version f
(current version g).
If it can read the sstables, why can't they be used to repair? Is there
anything I can do besides r
Hi All, sorry for taking so long to answer. I was away from the
internet.
>> Héctor, when you say "I have upgraded all my cluster to 0.8.1", from
> >> which version was
> >> that: 0.7.something or 0.8.0 ?
0.7.6-2 to 0.8.1
> This is the same behavior I reported in 2768 as Aaron referenced ...
> >
Hi all,
I have upgraded all my cluster to 0.8.1. Today one of the disks in one
of the nodes died. After replacing the disk I tried running repair, but
this message appears:
INFO [manual-repair-bdb4055a-d370-4d2a-a1dd-70a7e4fa60cf] 2011-06-30
20:36:25,085 AntiEntropyService.java (line 179) Exclud
Btw, if I restart the node, then it happily proceeds with compaction.
El jue, 23-06-2011 a las 10:02 +0200, Héctor Izquierdo Seliva escribió:
> Hi Aaron. Reverted back to 4-32. Did the flush but it did not trigger
> any minor compaction. Ran compact by hand, and it picked only two
>
;
> -
> Aaron Morton
> Freelance Cassandra Developer
> @aaronmorton
> http://www.thelastpickle.com
>
> On 23 Jun 2011, at 02:04, Héctor Izquierdo Seliva wrote:
>
> > Hi All. I set the compaction threshold at minimum 2, maximum 2 and try
> > to r
ut I
think this is not normal behaviour.
El lun, 13-06-2011 a las 10:32 -0500, Jonathan Ellis escribió:
> As Terje already said in this thread, the threshold is per bucket
> (group of similarly sized sstables) not per CF.
>
> 2011/6/13 Héctor Izquierdo Seliva :
> > I was already
This is what I use:
http://code.google.com/p/simple-cassandra-monitoring/
Disclaimer: I did it myself, don't expect too much :P
El jue, 16-06-2011 a las 19:35 +0300, Viktor Jevdokimov escribió:
> There's possibility to use command line JMX client with standard
> Zabbix agent to request JMX count
I was already way over the minimum. There were 12 sstables. Also, is
there any reason why scrub got stuck? I did not see anything in the
logs. Via jmx I saw that the scrubbed bytes were equal to one of the
sstables size, and it stuck there for a couple hours .
El lun, 13-06-2011 a las 22:55 +0900,
Hi All. I found a way to be able to compact. I have to call scrub on
the column family. Then scrub gets stuck forever. I restart the node,
and voila! I can compact again without any message about not having
enough space. This looks like a bug to me. What info would be needed to
fill a report? This
El vie, 10-06-2011 a las 23:40 +0900, Terje Marthinussen escribió:
> Yes, which is perfectly fine for a short time if all you want is to
> compact to one file for some reason.
>
>
> I run min_compaction_threshold = 2 on one system here with SSD. No
> problems with the more aggressive disk utiliza
El vie, 10-06-2011 a las 20:21 +0900, Terje Marthinussen escribió:
> bug in the 0.8.0 release version.
>
>
> Cassandra splits the sstables depending on size and tries to find (by
> default) at least 4 files of similar size.
>
>
> If it cannot find 4 files of similar size, it logs that message
Hi Terje,
There are 12 SSTables, so I don't think that's the problem. I will try
anyway and see what happens.
El vie, 10-06-2011 a las 20:21 +0900, Terje Marthinussen escribió:
> bug in the 0.8.0 release version.
>
>
>
> Cassandra splits the sstables depending on size and tries to find (by
>
Hi, I'm running a test node with 0.8, and everytime I try to do a major
compaction on one of the column families this message pops up. I have
plenty of space on disk for it and the sum of all the sstables is
smaller than the free capacity. Is there any way to force the
compaction?
I think I will follow the advice of better balancing and I will split
the index into several pieces. Thanks everybody for your input!
I'm actually using it in a couple of nodes, but is slower than directly
accesing the data in a ssd.
El jue, 09-06-2011 a las 11:10 -0400, Chris Burroughs escribió:
> On 06/08/2011 05:54 AM, Héctor Izquierdo Seliva wrote:
> > Is there a way to control what sstables go to what dat
El jue, 09-06-2011 a las 13:28 +0200, Richard Low escribió:
> Remember also that partitioning is done by rows, not columns. So
> large rows are stored on a single host. This means they can't be load
> balanced and also all requests to that row will hit one host. Having
> separate rows will allow
Hi,
I have an index I use to translate ids. I usually only read a column at
a time, and it's becoming a bottleneck. I could rewrite the application
to read a bunch at a time but it would make the application logic much
harder, as it would involve buffering incoming data.
As far as I know, to read
El mié, 08-06-2011 a las 08:42 -0500, Jonathan Ellis escribió:
> No. https://issues.apache.org/jira/browse/CASSANDRA-2749 is open to
> track this but nobody is working on it to my knowledge.
>
> Cassandra is fine with symlinks at the data directory level but I
> don't think that helps you, since y
Hi,
Is there a way to control what sstables go to what data directory? I
have a fast but space limited ssd, and a way slower raid, and i'd like
to put latency sensitive data into the ssd and leave the other data in
the raid. Is this possible? If not, how well does cassandra play with
symlinks?
Hi everyone. I see in the logs that Concurrent Mark Sweep is taking 12
seconds to do its stuff. Is this normal? There is no stop-the-world GC,
it just takes 12 seconds.
Configuration: 0.7.5 , 8GB Heap, 16GB machines. 7 * 64 MB memtables.
for RecentBloomFilterFalsePositives and
> BloomFilterFalsePositives the non ratio ones ?
>
> -
> Aaron Morton
> Freelance Cassandra Developer
> @aaronmorton
> http://www.thelastpickle.com
>
> On 11 May 2011, at 19:53, Héctor Izquierdo Seliva wrote:
>
> > El mié, 11-
El mié, 11-05-2011 a las 14:24 +1200, aaron morton escribió:
> What version and what were the values for RecentBloomFilterFalsePositives and
> BloomFilterFalsePositives ?
>
> The bloom filter metrics are updated in SSTableReader.getPosition() the only
> slightly odd thing I can see is that we do
El lun, 09-05-2011 a las 17:58 +0200, Peter Schuller escribió:
> > I have a few sstables with around 500 million keys, and memory usage has
> > grown a lot, I suppose because of the indexes. This sstables are
> > comprised of skinny rows, but a lot of them. Would tuning index interval
> > make the
Hi everyone.
I have a few sstables with around 500 million keys, and memory usage has
grown a lot, I suppose because of the indexes. This sstables are
comprised of skinny rows, but a lot of them. Would tuning index interval
make the memory usage go down? And what would the performance hit be?
I
>
> El mié, 04-05-2011 a las 21:02 +1200, aaron morton escribió:
> > Certainly sounds a bit sick.
> >
> > The first error looks like it happens when the index file points to the
> > wrong place in the data file for the SSTable. The second one happens when
> > the index file is corrupted. The
t helps.
>
> -
> Aaron Morton
> Freelance Cassandra Developer
> @aaronmorton
> http://www.thelastpickle.com
>
> On 4 May 2011, at 17:54, Héctor Izquierdo Seliva wrote:
>
> >
> > Hi Aaron
> >
> > It has no data files whatsoever.
t process did you use to recover ?
>
> Also what version are you running and what's the recent upgrade history ?
>
> Cheers
> Aaron
>
> On 3 May 2011, at 23:09, Héctor Izquierdo Seliva wrote:
>
> > Hi everyone. One of the nodes in my 6 node cluster died with disk
Hi everyone. One of the nodes in my 6 node cluster died with disk
failures. I have replaced the disks, and it's clean. It has the same
configuration (same ip, same token).
When I try to restart the node it starts to throw mmap underflow
exceptions till it closes again.
I tried setting io to stand
El mié, 20-04-2011 a las 23:00 +1200, aaron morton escribió:
> Looks like a bug, I've added a patch
> here https://issues.apache.org/jira/browse/CASSANDRA-2519
>
>
> Aaron
>
That was fast! Thanks Aaron
El mié, 20-04-2011 a las 09:08 +1200, aaron morton escribió:
> Yes, I saw that.
>
> Wanted to know what "issue deletes through pelops" means so I can work out
> what command it's sending to cassandra and hopefully I don't waste my time
> looking in the wrong place.
>
> Aaron
>
Oh, sorry. Di
I poste it a couple of messages back, but here it is again:
I'm using 0.7.4. I have a file with all the row keys I have to delete
(around 100 million) and I just go through the file and issue deletes
through pelops. Should I manually issue flushes with a cron every x
time?
I have a file with all the row keys I have to
> delete
> (around 100 million) and I just go through the file and issue
> deletes
> through pelops.
>
> Should I manually issue flushes with a cron every x time?
>
>
rows and my
> > insert rate is very low, so I don't have a lot of compactions. What
> > should I do in this case? Lower the major compaction threshold and
> > memtable_operations to some very low number?
> >
> > Thanks
> >
> > El mar, 19-04-2011 a
would be feasible to somehow store what
parts of the sstables are hot on shutdown and re read those parts of the
files on startup. Could it be done in a similar way to the work that's
being done on page migrations? What do you think?
Thanks for your time!
>
> On 20 Apr 2011, at 00:41,
should I do in this case? Lower the major compaction threshold and
memtable_operations to some very low number?
Thanks
El mar, 19-04-2011 a las 17:36 +0200, Héctor Izquierdo Seliva escribió:
> Hi everyone. I've configured in one of my column families
> memtable_operations = 0.02 and sta
Hi everyone. I've configured in one of my column families
memtable_operations = 0.02 and started deleting keys. I have already
deleted 54k, but there hasn't been any flush of the memtable. Memory
keeps pilling up and eventually nodes start to do stop-the-world GCs. Is
this the way this is supposed
Shouldn't the dynamic snitch take into account response times and ask a
slow node for less requests? It seems that at node startup, only a
handfull of requests arrive to the node and it keeps up well, but
there's moment where there's more than it can handle with a cold cache
and starts droping mess
ol that is
> > > backed by
> > > Apache Commons Pool (assuming you're using Cassandra 0.7). I'm
> > > not saying
> > > it's perfect but it's NOT sharing a connection over multiple
> > > threads.
> > > Dan Hendry mentioned tha
How difficult do you think this could be? I would be interested into
developing this if it's feasible.
El vie, 15-04-2011 a las 16:19 +0200, Peter Schuller escribió:
> > Hi everyone, is there any recommended procedure to warm up a node before
> > bringing it up?
>
> Currently the only out-of-the-
Hi everyone, is there any recommended procedure to warm up a node before
bringing it up?
Thanks!
I'm not sure. Are you runing it in the same host as the cassandra node?
El mar, 12-04-2011 a las 22:54 +0500, Ali Ahsan escribió:
> On 04/12/2011 10:42 PM, Héctor Izquierdo Seliva wrote:
> >
> > I forgot to mention it has been coded against 0.7.x, and I'm not sure it
>
ks2:cf2,etc
I forgot to mention it has been coded against 0.7.x, and I'm not sure it
will work on 0.6.x. I'll try to add support for both 0.6.x and the new
0.8.x version as soon as possible.
>
> On 04/12/2011 07:26 PM, Héctor Izquierdo Seliva wrote:
> > Hi everyone.
> &
Hi everyone.
Looking for ways to monitor cassandra with zabbix I could not found
anything that was really usable, till I found mention of a nice class by
smeet. I have based my modification upon his work and now I give it back
to the community.
Here's the project url:
http://code.google.com/p/si
Thanks Aaron!
El mar, 12-04-2011 a las 23:52 +1200, aaron morton escribió:
> Bug in the CLI, created /
> fixed https://issues.apache.org/jira/browse/CASSANDRA-2458
>
>
> use 70 for now.
>
>
> Thanks
> Aaron
>
>
> On 12 Apr 2011, at 20:46, Héctor Izq
Hi everyone.
I've changed the read repair chance of one of my column families from
cassandra-cli with the following entry:
update column family with read_repair_chance = 0.7
I expected to see in the server log
readRepairChance=0.7
Instead I saw this
readRepairChance=0.006999,
S
El mié, 06-04-2011 a las 21:04 -0500, Jonathan Ellis escribió:
> "out of sequence response" is thrift's way of saying "I got a response
> for request Y when I expected request X."
>
> my money is on using a single connection from multiple threads. don't do
> that.
>
I'm not using thrift direct
El mié, 06-04-2011 a las 21:04 -0500, Jonathan Ellis escribió:
> "out of sequence response" is thrift's way of saying "I got a response
> for request Y when I expected request X."
>
> my money is on using a single connection from multiple threads. don't do
> that.
I'm not using thrift direc
El mié, 06-04-2011 a las 09:18 +0200, Héctor Izquierdo Seliva escribió:
> I took a look at vmstats, and there was no swap. Also, our monitoring
> tools showed no swap being used at all. It's running with mlockall and
> all that. 8GB heap on a 16GB machine
>
I tried disabling swa
les fly
> visit digitalpigeon.com
>
> On Tuesday, 5 April 2011 at 11:43 PM, Héctor Izquierdo Seliva wrote:
>
> > El mar, 05-04-2011 a las 09:35 -0400, Dan Hendry escribió:
> > > I too have seen the out of sequence response problem. My solution
> > > has just been
I took a look at vmstats, and there was no swap. Also, our monitoring
tools showed no swap being used at all. It's running with mlockall and
all that. 8GB heap on a 16GB machine
El mar, 05-04-2011 a las 21:24 +0200, Peter Schuller escribió:
> > Would you recommend to disable system swap as a rule?
apping, and full GCs are taking
around 5 seconds. I'm running cassandra with a heap of 8 GB. Should I tune this
somehow?
Is any of this wrong?
>
> > > -----Original Message-
> > > From: Héctor Izquierdo Seliva [mailto:izquie...@strands.com]
> > > Sen
> -Original Message-
> > From: Héctor Izquierdo Seliva [mailto:izquie...@strands.com]
> > Sent: April-05-11 8:30
> > To: user@cassandra.apache.org
> > Subject: batch_mutate failed: out of sequence response
> >
> > Hi everyone. I'm having trouble while
your response!
Héctor
> -Original Message-
> From: Héctor Izquierdo Seliva [mailto:izquie...@strands.com]
> Sent: April-05-11 8:30
> To: user@cassandra.apache.org
> Subject: batch_mutate failed: out of sequence response
>
> Hi everyone. I'm having trouble while
Hi everyone. I'm having trouble while inserting big amounts of data into
cassandra. I'm getting this exception:
batch_mutate failed: out of sequence response
I'm gessing is due to very big mutates. I have made the batch mutates
smaller and it seems to be behaving. Can somebody shed some light?
T
still has to seek
to the position of the columns within the row.
How much space does the key cache uses per row? This would make the
number of rows increase by a big factor.
> On 22/02/2011, at 3:56 AM, Héctor Izquierdo Seliva
> wrote:
>
> > Hi Everyone.
> >
> >
ith options
= [{DC1:1,DC2:1}] and the same keyspace in DC2 with options [{DC2:1,
DC1:0}]. Is that correct?
> Aaron
>
> On 22/02/2011, at 3:10 AM, Héctor Izquierdo Seliva
> wrote:
>
> >
> > Hi all.
> >
> > Is there a way (besides changing the code) to repli
Hi Everyone.
I'm testing performance differences of millions of columns in a row vs
millions of rows. So far it seems wide rows perform better in terms of
reads, but there can be potentially hundreds of millions of columns in a
row. Is this going to be a problem? Should I go with individual rows?
Hi all.
Is there a way (besides changing the code) to replicate data from a Data
center 1 to a Data center 2, but not the other way around? I need to
have a preproduction environment with production data, and ideally with
only a fraction of the data (for example, by key preffixes). I have
poked a
Thanks! I totally overlooked that.
El lun, 21-02-2011 a las 08:14 +1300, Aaron Morton escribió:
> The best examples I know of are in the internal cli help, and
> conf/casandra.yaml
> Aaron
>
> On 19/02/2011, at 12:51 AM, Héctor Izquierdo Seliva
> wrote:
>
> > Hi!
Hi!
Can some body give me some hints about how to configure a keyspace with
NetworkTopologyStrategy via cassandra-cli? Or what is the preferred
method to do so?
Thanks!
Hi everyone.
I have a question about data modeling in my application. I have to store
items of a customer, and I can do it in one fat row per customer where
the column name is the id and the value a json serialized object, or one
entry per item with the same layout. This data is updated almost eve
no way - well I certainly feel stupid! Is this new, it worked without
> > it on beta 3?
> >
> > 2010/11/26 Héctor Izquierdo Seliva :
> >> Try ending the lines with ;
> >>
> >> Regards
> >>
> >> El vie, 26-11-2010 a las 21:25 +1100, jasonm
Try ending the lines with ;
Regards
El vie, 26-11-2010 a las 21:25 +1100, jasonmp...@gmail.com escribió:
> Hi,
>
> So I had this working perfectly with beta 3 and now it fails.
> Basically what I do is follows:
>
> 1) Extract new rc1 tarball.
> 2) Prepare location based on instructions in Readm
That happened to me too. Try with a ; at the end of the line.
El jue, 25-11-2010 a las 17:22 +, Marcin escribió:
> Hi guys,
>
> I am having weird problem, cassandra is working but can't get
> cassandra-cli to work.
>
> When I run command - any command like even help and hit error I am not
El lun, 11-10-2010 a las 11:08 -0400, Edward Capriolo escribió:
Inlined:
> 2010/10/11 Héctor Izquierdo Seliva :
> > Hi everyone.
> >
> > I'm sure this question or similar has come up before, but I can't find a
> > clear answer. I have to store a unknown nu
Hi everyone.
I'm sure this question or similar has come up before, but I can't find a
clear answer. I have to store a unknown number of items in cassandra,
which can vary from a few hundreds to a few millions per customer.
I read that in cassandra wide rows are better than a lot of rows, but
then
84 matches
Mail list logo