thank you very much for the reply. which brings me more confidence on
cassandra.
I will try the automation tools, the examples you've listed seems quite
promising!
about the decommission problem, here is the link:
http://cassandra-user-incubator-apache-org.3065146.n2.nabble.com/how-to-decommissio
As mentioned by Aaron, yes we run hundreds of Cassandra nodes across
multiple clusters. We run with RF of 2 and 3 (most common).
We use commodity hardware and see failure all the time at this scale. We've
never had 3 nodes that were in same replica set, fail all at once. We
mitigate risk by being
Hi,
Let's say if I want to migrate data from one cluster to another cluster, in
addition to snapshots, is there a need to also backup the commit log?
As far as I know, some of the data inside commit log might not have been
flushed to sstable during snapshots, therefore, if I only backup the
sstab
Can you please Watchitoo.com (its' free) and broadcast the event?
On Fri, Jul 8, 2011 at 8:54 PM, Richard Low wrote:
> Hi all,
>
> If you're in San Francisco for CassandraSF on Monday 11th, then come
> and join fellow Cassandra users and committers on Sunday evening.
> Starting at 6:30pm at Thir
On Sat, Jul 9, 2011 at 2:08 PM, Eldad Yamin wrote:
> Can you please Watchitoo.com (its' free) and broadcast the event?
>
> On Fri, Jul 8, 2011 at 8:54 PM, Richard Low wrote:
>
>> Hi all,
>>
>> If you're in San Francisco for CassandraSF on Monday 11th, then come
>> and join fellow Cassandra users
Flush is done as part of the snapshot, so if you're not doing writes
post-snapshot, there's no need to ship commitlogs. If you ARE still
doing writes, enable encremental backups to get sstables flushed
during the transfer.
On Sat, Jul 9, 2011 at 6:23 AM, Boris Yen wrote:
> Hi,
> Let's say if I w
>> - Have you been running repair consistently ?
>
> Nop, only when something breaks
This is unrelated to the problem you were asking about, but if you
never run delete, make sure you are aware of:
http://wiki.apache.org/cassandra/Operations#Frequency_of_nodetool_repair
http://wiki.apache.org/cas
Hi,
I have few questions:
*Secondary index*
1. Is there a limit on the number of columns in a single column family
that serve as secondary indexes?
2. Does performance decrease (significantly) if the uniqueness of the
column’s values is high?
*Twissandra*
1. Why in the source (o
Hi all,
we are currently looking at using Cassandra to store highly skewed RDF data.
With the indexes we use it may happen that a single row contains up to 20% of
the whole dataset, meaning that it can grow larger than available disk space on
single nodes. In [1], it says that this limitation i
Perhaps I misunderstand your proposal, but it seems that even with your
manual key placement schemes, the row would still be huge, no matter what
node it gets placed on. A better solution might be figuring out how to make
each row into a few smaller ones to get better balancing of load and also
fa
Hi,
On 09.07.2011, at 23:37, Dan Kuebrich wrote:
> Perhaps I misunderstand your proposal, but it seems that even with your
> manual key placement schemes, the row would still be huge, no matter what
> node it gets placed on. A better solution might be figuring out how to make
> each row into
Hi Peter.
I have a problem with repair, and it's that it always brings the node
doing the repairs down. I've tried setting index_interval to 5000, and
it still dies with OutOfMemory errors, or even worse, it generates
thousands of tiny sstables before dying.
I've tried like 20 repairs during thi
Check the log on all the machines for ERROR messages. An error on any of the
nodes could have caused the streaming to hang. nodetool netstats will let you
know if there is a failed stream.
AFAIK if you restart the cass service on 1 it will forget it was leaving and
rejoin in a normal state.
c
> about the decommission problem, here is the link:
> http://cassandra-user-incubator-apache-org.3065146.n2.nabble.com/how-to-decommission-two-slow-nodes-td5078455.html
The key part of that post is "and since the second node was under heavy load,
and not enough ram, it was busy GCing and worked
> Is there a limit on the number of columns in a single column family that
> serve as secondary indexes?
AFAIK there is no coded limit, however every index is implemented as another
(hidden) Column Family that inherits the settings of the parent CF. So under
0.7 you may run out of memory, under
> Nop, only when something breaks
Unless you've been working at QUORUM life is about to get trickier. Repair is
an essential part of running a cassandra cluster, without it you risk data loss
and dead data coming back to life.
If you have been writing at QUORUM, so have a reasonable expectatio
Sounds like your non-repair workload is using too much of the heap.
Alternatively, you could have a very large supercolumn that causes the
OOM when it is read.
2011/7/9 Héctor Izquierdo Seliva :
> Hi Peter.
>
> I have a problem with repair, and it's that it always brings the node
> doing the rep
Thanks. I will try it.
Boris
On Sat, Jul 9, 2011 at 8:44 PM, Jonathan Ellis wrote:
> Flush is done as part of the snapshot, so if you're not doing writes
> post-snapshot, there's no need to ship commitlogs. If you ARE still
> doing writes, enable encremental backups to get sstables flushed
> d
All the important stuff is using QUORUM. Normal operation uses around
3-4 GB of heap out of 6. I've also tried running repair on a per CF
basis, and still no luck. I've found it's faster to bootstrap a node
again than repairing it.
Once I have the cluster in a sane state I'll try running a repair
I missed the consistency level part, thanks very much for the explanation.
that is clear enough.
On Sun, Jul 10, 2011 at 7:57 AM, aaron morton wrote:
> about the decommission problem, here is the link:
> http://cassandra-user-incubator-apache-org.3065146.n2.nabble.com/how-to-decommission-two-slow
20 matches
Mail list logo