Thank you guys. It makes sense.
I'll have repair-pr schedule on each node.
On Thu, Oct 18, 2012 at 3:39 AM, aaron morton wrote:
> Without -pr the repair works on all token ranges the node is a replica
> for.
>
> With -pr it only repairs data in the token range it is assigned. In your
> case wh
On Thu, Oct 18, 2012 at 7:30 AM, Alain RODRIGUEZ wrote:
> Hi Jonathan.
>
> We are currently running the datastax AMI on amazon. Cassandra is in version
> 1.1.2.
>
> I guess that the datastax repo (deb http://debian.datastax.com/community
> stable main) will be updated directly in 1.1.6 ?
Yes.
>
Bryan , Manu
Can you contribute your experience to this ticket
https://issues.apache.org/jira/browse/CASSANDRA-4781 ? If you have some steps
to reproduce, or any information on schema changes you have made that would be
useful.
The ticket is against 1.2.0b1 - so please include
Cool.
If you get it again grab nodetool gossipinfo from a few machines.
Cheers
-
Aaron Morton
Freelance Developer
@aaronmorton
http://www.thelastpickle.com
On 19/10/2012, at 3:32 AM, Rene Kochen wrote:
> Thanks Aaron,
>
> Telnet works (in both directions).
>
> After a norm
> I replaced it with a new node, IP 10.16.128.197 and again token 0 with a
> "-Dcassandra.replace_token=0" at startup
Good Good.
How long ago did you bring the new node on ? There is a fail safe to remove
128.210 after 3 days if it does not gossip to other nodes.
I *thought* that remove_token
It depends if C* knows about the column.
You can check this by looking at the schema in cassandra-cli, see the online
help for show schema.
AFAIK most higher level libraries will get the data type info via the API. They
use this to determine the wire format for serialisation. Crack open the c
On Thu, Oct 18, 2012 at 2:31 PM, Jeremy Hanna
wrote:
>
> On Oct 18, 2012, at 3:52 PM, Andrey Ilinykh wrote:
>
>> On Thu, Oct 18, 2012 at 1:34 PM, Michael Kjellman
>> wrote:
>>> Not sure I understand your question (if there is one..)
>>>
>>> You are more than welcome to do CL ONE and assuming you
After some time. I believe this is correct. The "load" seems to be
correlated to compactions/number of files for keyspace/IO etc.
Thanks all!
Regards,
On Thu, Oct 18, 2012 at 9:35 PM, aaron morton wrote:
> At times of high load check the CPU % for the java service running C* to
> confirm C* is
On Oct 18, 2012, at 3:52 PM, Andrey Ilinykh wrote:
> On Thu, Oct 18, 2012 at 1:34 PM, Michael Kjellman
> wrote:
>> Not sure I understand your question (if there is one..)
>>
>> You are more than welcome to do CL ONE and assuming you have hadoop nodes
>> in the right places on your ring things
1. Yes, you can absolutely benefit from data locality, and the InputSplits
will theoretically schedule the map task on Cassandra+Hadoop nodes that
have the data locally. If your application doesn't require you to worry
about that one pesky row that should be local to that node (and that node
is res
I believe that reading with CL.ONE will still cause read repair to be run
(in the background) 'read_repair_chance' of the time.
-Bryan
On Thu, Oct 18, 2012 at 1:52 PM, Andrey Ilinykh wrote:
> On Thu, Oct 18, 2012 at 1:34 PM, Michael Kjellman
> wrote:
> > Not sure I understand your question (i
On Thu, Oct 18, 2012 at 1:34 PM, Michael Kjellman
wrote:
> Not sure I understand your question (if there is one..)
>
> You are more than welcome to do CL ONE and assuming you have hadoop nodes
> in the right places on your ring things could work out very nicely. If you
> need to guarantee that you
Not sure I understand your question (if there is one..)
You are more than welcome to do CL ONE and assuming you have hadoop nodes
in the right places on your ring things could work out very nicely. If you
need to guarantee that you have all the data in your job then you'll need
to use QUORUM.
If
On Thu, Oct 18, 2012 at 1:24 PM, Michael Kjellman
wrote:
> Well there is *some* data locality, it's just not guaranteed. My
> understanding (and someone correct me if I'm wrong) is that
> ColumnFamilyInputFormat implements InputSplit and the getLocations()
> method.
>
> http://hadoop.apache.org/do
Well there is *some* data locality, it's just not guaranteed. My
understanding (and someone correct me if I'm wrong) is that
ColumnFamilyInputFormat implements InputSplit and the getLocations()
method.
http://hadoop.apache.org/docs/mapreduce/current/api/org/apache/hadoop/mapre
duce/InputSplit.html
On Thu, Oct 18, 2012 at 12:00 PM, Michael Kjellman
wrote:
> Unless you have Brisk (however as far as I know there was one fork that got
> it working on 1.0 but nothing for 1.1 and is not being actively maintained
> by Datastax) or go with CFS (which comes with DSE) you are not guaranteed
> all dat
Honestly, I think what they did re Brisk development is fair. They left
the code for any of us in the community to improve it and make it
compatible with newer versions and they need to make money as a company as
well. They already contribute so much to the Cassandra community in
general and they a
I am surprise that it was abandoned this way. So if I want to use
Brisk on Cassandra 1.1 I have to use DataStax Entreprise service...
On Thu, Oct 18, 2012 at 3:00 PM, Michael Kjellman
wrote:
>
> Unless you have Brisk (however as far as I know there was one fork that
> got it working on 1.0 but no
Unless you have Brisk (however as far as I know there was one fork that got it
working on 1.0 but nothing for 1.1 and is not being actively maintained by
Datastax) or go with CFS (which comes with DSE) you are not guaranteed all data
is on that hadoop node. You can take a look at the forks if in
A recent thread made it sound like Brisk was no longer a datastax supported
thing (it's DataStax Enterpise, or DSE, now):
http://www.mail-archive.com/user@cassandra.apache.org/msg24921.html
In particular this response:
http://www.mail-archive.com/user@cassandra.apache.org/msg25061.html
On Thu, Oc
Why don't you look into Brisk:
http://www.datastax.com/docs/0.8/brisk/about_brisk
On Thu, Oct 18, 2012 at 2:46 PM, Andrey Ilinykh wrote:
> Hello, everybody!
> I'm thinking about running hadoop jobs on the top of the cassandra
> cluster. My understanding is - hadoop jobs read data from local node
Hello, everybody!
I'm thinking about running hadoop jobs on the top of the cassandra
cluster. My understanding is - hadoop jobs read data from local nodes
only. Does it mean the consistency level is always ONE?
Thank you,
Andrey
Hi Steve,
Also confirming this. After having a node go down on Cassandra 1.0.8
there seems to be hinted handoff between two of our 4 nodes every 10
minutes. Our setup also shows 0 rows. It does not appear to have any
effect on the operation of the ring, just fills up the log files.
- David
On
I installed Cassandra on three nodes. I then ran a test suite against them to
generate load. The test suite is designed to generate the same type of load
that we plan to have in production. As one of many tests, I reset one of the
nodes to check the failure/recovery modes. Cassandra worked just
In a 4 node cluster running Cassandra 1.1.5 with sun jvm 1.6.0_29-b11
(64-bit), the nodes are often getting "stuck" in state where CMS
collections of the old space are constantly running.
The JVM configuration is using the standard settings in cassandra-env --
relevant settings are included below.
I think I am seeing the same issue, but it doesn't seem to be related to the
schema_columns. I understand that repair is supposed to be intensive, but this
is bringing the associated machine to its knees, to the point that logging on
the machine takes a very, very long time and requests are no
Thanks Aaron,
Telnet works (in both directions).
After a normal (i.e. without discarding ring state) restart of the node
reporting the other one as down, the ring shows "up" again. So a node
restarts fixes the incorrect state.
I see this error occasionally.
I will further investigate and post m
Hi all
I'm running Cassandra 1.0.11 on Ubuntu 11.10.
I've got a ghost node which keeps showing up on my ring.
A node living on IP 10.16.128.210 and token 0 died and had to be replaced.
I replaced it with a new node, IP 10.16.128.197 and again token 0 with a
"-Dcassandra.replace_token=0" at start
This is specifically why Cassandra and even PlayOrm are going the
direction of "partial schemas". Everything in cassandra in raw form is
just bytes. If you don't tell it the types, it doesn't know how to
translate it. PlayOrm and other ORM layers are the same way though in
these noSQL ORMs you t
Hi Jonathan.
We are currently running the datastax AMI on amazon. Cassandra is in
version 1.1.2.
I guess that the datastax repo (deb
http://debian.datastax.com/communitystable main) will be updated
directly in 1.1.6 ?
"Replaying already-flushed data a second time is harmless -- except
for counte
Hi all,
I am wondering if there is a way to know the column type of an already stored
value in Cassandra.
My specific case is to get a column value of a known column name but not type.
greetings
Ambes
> Follow up question: Is it safe to abort the compactions happening after node
> repair?
It is always safe to abort a compaction. The purpose of compaction is to
replicate the current truth in a more compact format. It does not modify data,
it just creates new files. The worse case would be kill
Very slim reason to link to my favourite Joe Celko
(http://en.wikipedia.org/wiki/Joe_Celko) quote:
"
'LOL! My wife is an ordained Soto Zen priest. I would say after 30 years
together, I'd go with her. She is the only person who understood NULLs
immediately.
http://www.simple-talk.com/opinion/g
>> Yes, i understand that. Reason why i am asking is, with this i need to split
>> them to get actual column name using ":" as a seperator.
The ":" is a artefact of the cassandra-cli, nothing something you will have to
deal with via the thrift API. Internally we do not store the values with ":"
s
Without -pr the repair works on all token ranges the node is a replica for.
With -pr it only repairs data in the token range it is assigned. In your case
when you ran it on node 0 with RF the token range form node 0 was repaired on
nodes 0, 1 and 2. The other token ranges on nodes 0, 1 and 2 w
At times of high load check the CPU % for the java service running C* to
confirm C* is the source of load.
If the load is generated from C* check the logs (or use OpsCentre / other
monitoring) to see if it correlated to compaction, or Garbage Collection or
repair or high throughput.
Cheers
You can double check the node reporting 9.109 as down can telnet to port 7000
on 9.109.
Then I would restart 9.109 with -Dcassandra.load_ring_state=false added as a
JVM param in cassandra-env.sh.
If is still shows as down can you post the output from nodetool gossipinfo from
9.109 and the no
I have a four node EC2 cluster.
Three machines show via nodetool ring that all machines are UP.
One machine shows via nodetool ring that one machine is DOWN.
If I take a closer to the machine reporting the other machine as down, I
see the following:
- StorageService.UnreachableNodes = 10.49.9.10
38 matches
Mail list logo