I tried it on 1 column family. I believe there is a bug in 0.8* where
repair ignores the cf. I tried this multiple times on different nodes.
Every time the disk util was going uo to 80% on a 500 GB disk. I would
eventually kill the repair. I only have 60GB worth data. I see this JIRA -
https://iss
big buffer size will use more Heap memory at creation of the tables.
Not sure impact on server side, but shouldn't be a big difference. I
personally use 512Mb.
2012/4/28 sj.climber :
> Can anyone comment on best practices for setting the buffer size used by
> SSTableSimpleUnsortedWriter? I'm
About a year ago I started getting a strange feeling that
the noSQL community is busy re-creating RDBMS in minute detail.
Why did we bother in the first place?
Maxim
On 4/27/2012 6:49 PM, Data Craftsman wrote:
> Howdy,
>
> Some Polyglot Persistence(NoSQL) products started support server side
>
Tamar
Please don't jump to other users discussions. If you want to ask any issue,
create a new one, please.
Thanks.
--
View this message in context:
http://cassandra-user-incubator-apache-org.3065146.n2.nabble.com/Cassandra-backup-question-regarding-commitlogs-tp7508823p7511913.html
Sent from
Hi
Currently I am taking daily snapshot on my keyspace in production and
already enable the incremental backups as well.
According to the documentation, the incremental backup option will create an
hard-link to the backup folder when new sstable is flushed. Snapshot will
copy all the data/index/e
Jason,
I'm using plenty of secondary indexes with no problem at all.
Looking at your example,as I think you understand, you forgo indexes by
combining two conditions in one query, thinking along the lines of what is
often done in RDBMS. A scan is expected in this case, and there is no
magic to av
Did you get a solution on this one ?
It looks like you ran out of memory on the machine…
> Caused by: java.lang.OutOfMemoryError: Map failed
> at sun.nio.ch.FileChannelImpl.map0(Native Method)
> ... 7 more
cheers
-
Aaron Morton
Freelance Developer
@aaronmorton
h
Check there is a single schema version on the cluster, in the cassandra-cli use
describe cluster;
Cheers
-
Aaron Morton
Freelance Developer
@aaronmorton
http://www.thelastpickle.com
On 26/04/2012, at 3:33 AM, mdione@orange.com wrote:
> De : mdione@orange.com [mailto:md
> node that fail had the token id of 0 (this is the seed node - right?).
Seed nodes are listed in the seeds: section of the cassandra.yaml file.
Using 0 as a token for a node is normal.
Cheers
-
Aaron Morton
Freelance Developer
@aaronmorton
http://www.thelastpickle.com
On 26/04
Also i would avoid using HaProxy is possible. The best judge of a nodes
availability is the client, and it can varies per row key.
The exception is when you are using a web server that does not support state,
such as php. The solution is not to use php.
Cheers
-
Aaron Morton
F
Depends on your definition of significantly, there are a few things to
consider.
* Reading from SSTables for a request is a serial operation. Reading from 2
SSTables will take twice as long as 1.
* If the data in the One Big File™ has been overwritten, reading it is a waste
of time. And it w
> . Is there a better way to solve this in real time.
Not really. If however you can send a row level delete before the insert you
dont need to read first. Of course that deletes all the other data :)
If you create a secondary index on a column value, the index will be updated
when you change th
Message Queue is often not a great use case for Cassandra. For information on
how to handle high delete workloads see
http://www.datastax.com/dev/blog/leveled-compaction-in-apache-cassandra
It hard to create a model without some idea of the data load, but I would
suggest you start with:
CF: Us
Did you restart ? All good?
Cheers
-
Aaron Morton
Freelance Developer
@aaronmorton
http://www.thelastpickle.com
On 27/04/2012, at 9:49 AM, Bryce Godfrey wrote:
> This is the second node I’ve joined to my cluster in the last few days, and
> so far both have become stuck at 100%
Looks a bit like https://issues.apache.org/jira/browse/CASSANDRA-3579 but that
was fixed in 1.0.7
Is this still an issue ? Are you able to reproduce the fault ?
Cheers
-
Aaron Morton
Freelance Developer
@aaronmorton
http://www.thelastpickle.com
On 27/04/2012, at 6:56 PM, Patr
That sounds right to me.
A
-
Aaron Morton
Freelance Developer
@aaronmorton
http://www.thelastpickle.com
On 29/04/2012, at 5:00 AM, Paolo Bernardi wrote:
> Apparently IntegerType is based on Java's BigInteger.
>
> http://git-wip-us.apache.org/repos/asf?p=cassandra.git;a=blob_pla
When you start a node does it log that it's opening SSTables ?
After starting what does nodetool cfstats say for the node ?
Can you connect with cassandra-cli and do a get ?
Cheers
-
Aaron Morton
Freelance Developer
@aaronmorton
http://www.thelastpickle.com
On 29/04/2012, at 10
Each mutation is applied to the commit log before being applied to the
memtable. On server start the SSTables are read before replaying the commit
logs. This is part of the crash only software design and happens for every
start.
AFAIk there is no facility to snapshot commit log files as they ar
> 1. If I already have a Cassandra cluster running, would changing the
> incremental_backups parameter in the cassandra.yaml of each node, and then
> restart it do the trick?
Yes it is a per node setting.
> 2. Assuming I am creating a daily snapshot, what is the gain from setting
> incrementa
Hi!
I wonder what are the advantages of doing incremental snapshot over non
incremental?
Are the snapshots smaller is size? Are there any other implications?
Thanks,
*Tamar Fraenkel *
Senior Software Engineer, TOK Media
[image: Inline image 1]
ta...@tok-media.com
Tel: +972 2 6409736
Mob: +972
Hi Aaron
Thanks for the comments. Yes for the durability will keep them in a safe
place. But such crash situation, how can I restore the data (because those
are not in a SSTable and only in commit log).
Do I need to replay only that commit log when server starts after crash?
Will it override the
Hello List,
I need suggestion/ recommendation on time series data.
I have requirement where users belongs to different timezone and they can
subscribe to global group.
When users at specific timezone send update to group it is available to
every user in different timezone.
I am using GroupSubscr
22 matches
Mail list logo