Re: How many Column Families can Cassandra handle?

2013-09-30 Thread Aaron Morton
At 600 CF's I would expect to see very frequent flushing to disk, as the 
algorithm that drives this from a memory standpoint is sensitive to the number 
of CF's. 

Additionally, and from experience on earlier versions, you can expect it to 
take over half an hour to make schema changes to over 500 CFs. 

Finally, if you ever have performance problems it's a lot harder to diagnose on 
a system with 600 CF's that it is one with 60. 

Hope that helps. 

-
Aaron Morton
New Zealand
@aaronmorton

Co-Founder & Principal Consultant
Apache Cassandra Consulting
http://www.thelastpickle.com

On 27/09/2013, at 7:05 AM, Krishna Pisupat  wrote:

> I don't know the full use case. However, for a generic time series scenario, 
> we can make the timestamp (may be unto second part) part of the key and write 
> all the data into the same CF(one CF for all data). Again, it may not make 
> sense in your case, given the full use case. Just my 2 cents. 
> 
> 
> Thanks and Regards,
> Krishna Pisupat
> krishna.pisu...@gmail.com
> 
> 
> 
> On Sep 26, 2013, at 11:18 AM, "Hiller, Dean"  wrote:
> 
>> 600 is probably doable but each CF takes up memory……PlayOrm goes with a 
>> strategy that can virtualize CF's into one CF allowing less memory usage….we 
>> have 80,000 virtual CF's in cassandra through playorm….you can copy 
>> playorm's pattern if desired.  But 600 is probably doable but high.  10,000 
>> is not very doable.
>> 
>> But you would have to try out 600 to see if it works for you….it may not 
>> work…try and find out in your load and context.
>> 
>> NOTE: We have changed the 80,000 virtual CF's such that are in 10 real CF's 
>> these days so we get more parallel compaction going on.
>> 
>> Dean
>> 
>> From: Raihan Jamal mailto:jamalrai...@gmail.com>>
>> Reply-To: "user@cassandra.apache.org" 
>> mailto:user@cassandra.apache.org>>
>> Date: Thursday, September 26, 2013 11:39 AM
>> To: "user@cassandra.apache.org" 
>> mailto:user@cassandra.apache.org>>
>> Subject: How many Column Families can Cassandra handle?
>> 
>> I am working on a use case for Timeline series data. I have been told to 
>> create 600 column families in Cassandra. Meaning for 10 minutes, I will be 
>> having column families in Cassandra. Each second will have its own column 
>> family, so till 10 minutes which is 600 second, I will be having 600 column 
>> families...
>> 
>> In each second, we will write into that particular second column family.. so 
>> at 10 minutes (which is 600 second), we will write into 600 second column 
>> family..
>> 
>> I am wondering whether Cassandra will be able to handle 600 column families 
>> or not.. Right now, I am not sure how much data each column family will 
>> have... What I know so far is write will be coming at a rate of 20,000 
>> writes per second...
>> 
>> Can anyone shed some light into this?
> 



Re: Query about class org.apache.cassandra.io.sstable.SSTableSimpleWriter

2013-09-30 Thread Aaron Morton
> Thanks for the reply. Isn't the addColumn(IColumn col) method in the writer 
> private though?
> 
> 

Yes but I thought you had it in your examples, was included for completeness. 
use the official overloads. 

Cheers

-
Aaron Morton
New Zealand
@aaronmorton

Co-Founder & Principal Consultant
Apache Cassandra Consulting
http://www.thelastpickle.com

On 27/09/2013, at 4:12 PM, Jayadev Jayaraman  wrote:

> Thanks for the reply. Isn't the addColumn(IColumn col) method in the writer 
> private though? I know what to do now in order to construct a column with a 
> TTL now. Thanks.
> 
> On Sep 26, 2013 9:00 PM, "Aaron Morton"  wrote:
> > org.apache.cassandra.thrift.Column column; // initialize this with name, 
> > value, timestamp, TTL
> This is the wrong object to use.
> 
> one overload of addColumn() accepts IColumn which is from 
> org.apache.cassanda.db . The thrift classes are only use for the thrift API.
> 
> > What is the difference between calling writer.addColumn() on the column's 
> > name, value and timestamp, and writer.addExpiringColumn() on the column's 
> > name, value, TTL, timestamp and expiration timestamp ?
> They both add an column to the row. addExpiringColumn() adds an expiring 
> column, and addColumn adds a normal one.
> 
> only addExpiringColumn accepts a TTL (in seconds) for the column.
> 
> 
> > Does the former result in the column expiring still , in cassandra 1.2.x 
> > (i.e. does setting the TTL on a Column object change the name or value in a 
> > way so as to ensure the column will expire as required) ?
> No.
> An expiring column must be an ExpiringColumn column instance.
> The base IColumn interface does not have a TTL, only expiring columns do.
> 
> >  If not , what is the TTL attribute used for in the Column object ?
> The org.apache.cassandra.db.Column class does not have a TTL.
> 
> Cheers
> 
> 
> -
> Aaron Morton
> New Zealand
> @aaronmorton
> 
> Co-Founder & Principal Consultant
> Apache Cassandra Consulting
> http://www.thelastpickle.com
> 
> On 26/09/2013, at 12:44 AM, Jayadev Jayaraman  wrote:
> 
> > Can someone answer this doubt reg. SSTableSimpleWriter ? I'd asked about 
> > this earlier but it probably missed. Apologies for repeating the question 
> > (with minor additions)  :
> >
> > """
> > Let's say I've initialized a SSTableSimpleWriter instance and a new column 
> > with TTL set :
> >
> > org.apache.cassandra.io.sstable.SSTableSimpleWriter writer = new 
> > SSTableSimpleWriter( ... /* params here */);
> > org.apache.cassandra.thrift.Column column; // initialize this with name, 
> > value, timestamp, TTL
> >
> > What is the difference between calling writer.addColumn() on the column's 
> > name, value and timestamp, and writer.addExpiringColumn() on the column's 
> > name, value, TTL, timestamp and expiration timestamp ? Does the former 
> > result in the column expiring still , in cassandra 1.2.x (i.e. does setting 
> > the TTL on a Column object change the name or value in a way so as to 
> > ensure the column will expire as required) ? If not , what is the TTL 
> > attribute used for in the Column object ?
> > """
> >
> > Thanks,
> > Jayadev
> >
> >
> > On Tue, Sep 24, 2013 at 2:48 PM, Jayadev Jayaraman  
> > wrote:
> > Let's say I've initialized a SSTableSimpleWriter instance and a new column 
> > with TTL set :
> >
> > SSTableSimpleWriter writer = new SSTableSimpleWriter( ... /* params here 
> > */);
> > Column column;
> >
> > What is the difference between calling writer.addColumn() on the column's 
> > name and value, and writer.addExpiringColumn() on the column and its TTL ? 
> > Does the former result in the column expiring still , in cassandra 1.2.x ? 
> > Or does it not ?
> >
> >
> >
> 



Re: Connecting to a remote cassandra node..

2013-09-30 Thread Aaron Morton
See the comments for rpc_address and listen_address in the yaml file. These 
control which interfaces the server binds to. 

If you leave them as blank and DNS is correctly setup they will bind to the 
correct address. 

If you set them to 0.0.0.0 they will bind to all interfaces, you should not do 
this in production. 

Cheers

-
Aaron Morton
New Zealand
@aaronmorton

Co-Founder & Principal Consultant
Apache Cassandra Consulting
http://www.thelastpickle.com

On 27/09/2013, at 8:27 PM, Krishna Chaitanya  wrote:

> Thank you...but cant we dynamically connect to any system instead of changing 
> the .yaml file each time?? I want to take the input from the user regarding 
> the host and port that he wants to use...Thanks.
> 
> On Sep 27, 2013 1:38 PM, "Kais Ahmed"  wrote:
> hello,
> 
> you have to check listen_address in cassandra.yaml, change the localhost 
> value by the ip of the machine and restart cassandra  
> 
> 
> 2013/9/27 Krishna Chaitanya 
> Hello,
>   I am relatively new to cassandra. I am using a library called 
> libQtCassandra for accesing the cassandra database from my c++ programs. When 
> I try to connect to the localhost cassandra , everything seems fine but when 
> I try to connect to a remote node on which cassandra is up and running , it 
> says connection refused. Any help would be of a great value. Thank You...
> 
> -- 
> Regards,
> BNSK.
> 



Re: 2.0.1 counter replicate on write error

2013-09-30 Thread Aaron Morton
> ERROR [ReplicateOnWriteStage:19] 2013-09-27 10:17:14,778 CassandraDaemon.java 
> (line 185) Exception in thread Thread[ReplicateOnWriteStage:19,5,main]
> java.lang.AssertionError: DecoratedKey(-1754949563326053382, 
> a414b0c07f0547f8a75410555716ced6) != DecoratedKey(-1754949563326053382, 
> aeadcec8184445d4ab631ef4250927d0) in 
> /disk3/cassandra/data/struqrealtime/counters/struqrealtime-counters-jb-831953-Data.db
> at 
> org.apache.cassandra.db.columniterator.SSTableNamesIterator.read(SSTableNamesIterator.java:114)
> at 
> org.apache.cassandra.db.columniterator.SSTableNamesIterator.(SSTableNamesIterator.java:62)

When reading from an SSTable the position returned from the -Index.db / KEYS 
cache pointed to a row in the -Data.db component that was for a different row. 

> DecoratedKey(-1754949563326053382, aeadcec8184445d4ab631ef4250927d0)
Was what we were searching for

> DecoratedKey(-1754949563326053382, a414b0c07f0547f8a75410555716ced6)
Is what was found in the data component. 

The first part is the Token (M3 hash) the second is the key. It looks like a 
collision, but it could also be a bug somewhere else. 

Code in SSTableReader.getPosition() points to 
https://issues.apache.org/jira/browse/CASSANDRA-4687 and adds an assertion that 
is only trigger if TRACE logging is running. Can you add to the 4687 ticket and 
update the thread ? 

Cheers

-
Aaron Morton
New Zealand
@aaronmorton

Co-Founder & Principal Consultant
Apache Cassandra Consulting
http://www.thelastpickle.com

On 27/09/2013, at 10:50 PM, Christopher Wirt  wrote:

> Hello,
>  
> I’ve started to see a slightly worrying error appear in our logs 
> occasionally. We’re writing at 400qps per machine and I only see this appear 
> every 5-10minutes.
>  
> Seems to have started when I switched us to using the hsha thrift server this 
> morning. We’ve been running 2.0.1 ran off the sync thrift server since 
> yesterday without seeing this error.  But might not be related.
>  
> There are some machines in another DC still running 1.2.10.
>  
> Anyone seen this before? Have any insight?
>  
> ERROR [ReplicateOnWriteStage:19] 2013-09-27 10:17:14,778 CassandraDaemon.java 
> (line 185) Exception in thread Thread[ReplicateOnWriteStage:19,5,main]
> java.lang.AssertionError: DecoratedKey(-1754949563326053382, 
> a414b0c07f0547f8a75410555716ced6) != DecoratedKey(-1754949563326053382, 
> aeadcec8184445d4ab631ef4250927d0) in 
> /disk3/cassandra/data/struqrealtime/counters/struqrealtime-counters-jb-831953-Data.db
> at 
> org.apache.cassandra.db.columniterator.SSTableNamesIterator.read(SSTableNamesIterator.java:114)
> at 
> org.apache.cassandra.db.columniterator.SSTableNamesIterator.(SSTableNamesIterator.java:62)
> at 
> org.apache.cassandra.db.filter.NamesQueryFilter.getSSTableColumnIterator(NamesQueryFilter.java:87)
> at 
> org.apache.cassandra.db.filter.QueryFilter.getSSTableColumnIterator(QueryFilter.java:62)
> at 
> org.apache.cassandra.db.CollationController.collectAllData(CollationController.java:249)
> at 
> org.apache.cassandra.db.CollationController.getTopLevelColumns(CollationController.java:53)
> at 
> org.apache.cassandra.db.ColumnFamilyStore.getTopLevelColumns(ColumnFamilyStore.java:1468)
> at 
> org.apache.cassandra.db.ColumnFamilyStore.getColumnFamily(ColumnFamilyStore.java:1294)
> at org.apache.cassandra.db.Keyspace.getRow(Keyspace.java:332)
> at 
> org.apache.cassandra.db.SliceByNamesReadCommand.getRow(SliceByNamesReadCommand.java:55)
> at 
> org.apache.cassandra.db.CounterMutation.makeReplicationMutation(CounterMutation.java:100)
> at 
> org.apache.cassandra.service.StorageProxy$8$1.runMayThrow(StorageProxy.java:1107)
> at 
> org.apache.cassandra.service.StorageProxy$DroppableRunnable.run(StorageProxy.java:1897)
> at 
> java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1145)
> at 
> java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:615)
> at java.lang.Thread.run(Thread.java:724)



Re: What is the best way to install & upgrade Cassandra on Ubuntu ?

2013-09-30 Thread Aaron Morton
> I am not sure if I should use datastax's DSC or official Debian packages from 
> Cassandra. How do I choose between them for a production server ?
They are technically the same. 
The DSC update will come out a little after the Apache release, and I _think_ 
they release for every Apache release.

>  1.  when I upgrade to a newer version, would that retain my previous 
> configurations so that I don't need to configure everything again ? 

Yes if you select that when doing the package install. 

> 2.  would that smoothly replace the previous installation by itself ?

Yes

> 3.  what's the way (kindly, if you can tell the command) to upgrade ?


http://www.datastax.com/documentation/cassandra/2.0/webhelp/index.html#upgrade/upgradeC_c.html#concept_ds_yqj_5xr_ck

> 4. when should I prefer datastax's dsc to that ? (I need to install for 
> production env.)

Above

Hope that helps. 


-
Aaron Morton
New Zealand
@aaronmorton

Co-Founder & Principal Consultant
Apache Cassandra Consulting
http://www.thelastpickle.com

On 27/09/2013, at 11:01 PM, Ertio Lew  wrote:

> I am not sure if I should use datastax's DSC or official Debian packages from 
> Cassandra. How do I choose between them for a production server ?
> 
> 
> 
> On Fri, Sep 27, 2013 at 11:02 AM, Ertio Lew  wrote:
> 
>  Could you please clarify that:
> 1.  when I upgrade to a newer version, would that retain my previous 
> configurations so that I don't need to configure everything again ? 
> 2.  would that smoothly replace the previous installation by itself ?
> 3.  what's the way (kindly, if you can tell the command) to upgrade ?
> 4. when should I prefer datastax's dsc to that ? (I need to install for 
> production env.)
> 
> 
> On Fri, Sep 27, 2013 at 12:50 AM, Robert Coli  wrote:
> On Thu, Sep 26, 2013 at 12:05 PM, Ertio Lew  wrote:
> How do you install Cassandra on Ubuntu & later how do you upgrade the 
> installation on the node when an update has arrived ? Do you simply download 
> & replace the latest tar.gz, untar it to replace the older cassandra files? 
> How do you do it ? How does this upgrade process differ for a major version 
> upgrade, like say switching from 1.2 series to 2.0 series ?
> 
> Use the deb packages. To upgrade, install the new package. Only upgrade a 
> single major version. and be sure to consult NEWS.txt for any upgrade caveats.
> 
> Also be aware of this sub-optimal behavior of the debian packages :
> 
> https://issues.apache.org/jira/browse/CASSANDRA-2356
> 
> =Rob
> 
> 
> 



Re: HintedHandoff process does not finish

2013-09-30 Thread Aaron Morton
> What can be the reason for the handoff process not to finish?
Check for other errors about timing out during hint reply. 

> What would be the best way to recover from this situation?
If they are really causing trouble drop the hints via HintedHandoffManager JMX 
MBean or stopping the node and deleting the files on disk. Then use repair 
later. 

> What can be done to prevent this from happening again?
Hints are stored when either the node is down before the request starts or when 
the coordinator times out waiting for the remote node. Check the logs for nodes 
going down, and check the MessagingService MBean for TimedOuts from other 
nodes. This may indicate issues with a cross DC connection. 

Cheers

-
Aaron Morton
New Zealand
@aaronmorton

Co-Founder & Principal Consultant
Apache Cassandra Consulting
http://www.thelastpickle.com

On 27/09/2013, at 11:18 PM, Tom van den Berge  wrote:

> Hi,
> 
> One one of my nodes, the (storage) load increased dramatically (doubled), 
> within one or two hours. The hints column family was causing the growth. I 
> noticed one HintedHandoff process that was started some two hours ago, but 
> hadn't finished. Normally, these processes take only a few seconds, 15 
> seconds max, in my cluster.
> 
> The not-finishing process was handing the hints over to a host in another 
> data center. There were no warning or error messages in the logs, other than 
> the repeated "flushing high-traffic column family hints".
> I'm using Cassandra 1.2.3.
> What can be the reason for the handoff process not to finish?
> What would be the best way to recover from this situation?
> What can be done to prevent this from happening again?
> 
> Thanks in advance,
> Tom
> 
> 
> 
> 
> 



Re: temporal solution to CASSANDRA-5543: BUILD FAILED at gen-cql2-grammar target

2013-09-30 Thread Aaron Morton
It's an error in the antlr compilation, check the antlr versions. 

Cheers

-
Aaron Morton
New Zealand
@aaronmorton

Co-Founder & Principal Consultant
Apache Cassandra Consulting
http://www.thelastpickle.com

On 27/09/2013, at 11:53 PM, Miguel Angel Martin junquera 
 wrote:

> 
> 
> 
> 
> hi all:
> 
> Environment
> 
>   • apache-cassandra-2.0.1-src
>   • EC2
>   • Linux version 3.2.30-49.59.amzn1.x86_64 
> (mockbuild@gobi-build-31003) (gcc version 4.4.6 20110731 (Red Hat 4.4.6-3) 
> 
> When i try to build apache-cassandra-2.0.1-src in EC2 red had AMI, I  have 
> this error at  the target, gen-cql2-grammar:
> 
> 
> gen-cql2-grammar:
>  [echo] Building Grammar 
> /home/ec2-user/apache-cassandra/src/java/org/apache/cassandra/cql/Cql.g
> 
> ….
>
>   [java] warning(209): 
> /home/ec2-user/apache-cassandra/src/java/org/apache/cassandra/cql/Cql.g:638:1:
>  Multiple token rules can match input such as "'0'..'9'": INTEGER, FLOAT, UUID
>  [java] 
>  [java] As a result, token(s) FLOAT,UUID were disabled for that input
>  [java] warning(209): 
> /home/ec2-user/apache-cassandra/src/java/org/apache/cassandra/cql/Cql.g:634:1:
>  Multiple token rules can match input such as "'I'": K_INSERT, K_IN, K_INDEX, 
> K_INTO, IDENT, COMPIDENT
>  [java] 
>  [java] As a result, token(s) K_IN,K_INDEX,K_INTO,IDENT,COMPIDENT were 
> disabled for that input
>  [java] warning(209): 
> /home/ec2-user/apache-cassandra/src/java/org/apache/cassandra/cql/Cql.g:634:1:
>  Multiple token rules can match input such as "{'R', 'r'}": K_REVERSED, 
> IDENT, COMPIDENT
>  [java] 
>  [java] As a result, token(s) IDENT,COMPIDENT were disabled for that input
>  [java] warning(209): 
> /home/ec2-user/apache-cassandra/src/java/org/apache/cassandra/cql/Cql.g:634:1:
>  Multiple token rules can match input such as "'T'": K_LEVEL, K_TRUNCATE, 
> K_COLUMNFAMILY, K_TIMESTAMP, K_TTL, K_TYPE, IDENT, COMPIDENT
>  [java] 
>  [java] As a result, token(s) 
> K_TRUNCATE,K_COLUMNFAMILY,K_TIMESTAMP,K_TTL,K_TYPE,IDENT,COMPIDENT were 
> disabled for that input
>  [java] error(208): 
> /home/ec2-user/apache-cassandra/src/java/org/apache/cassandra/cql/Cql.g:654:1:
>  The following token definitions can never be matched because prior tokens 
> match the same input: 
> T__93,T__94,T__97,T__98,T__101,T__105,T__107,K_WITH,K_USING,K_USE,K_FIRST,K_COUNT,K_SET,K_APPLY,K_BATCH,K_TRUNCATE,K_IN,K_CREATE,K_KEYSPACE,K_COLUMNFAMILY,K_INDEX,K_ON,K_DROP,K_INTO,K_TIMESTAMP,K_TTL,K_ALTER,K_ADD,K_TYPE,RANGEOP,FLOAT,COMPIDENT,UUID,MULTILINE_COMMENT
> 
> BUILD FAILED
> /home/ec2-user/apache-cassandra/build.xml:218: Java returned: 1
> 
> 
> 
> 
> 
> 
> If  I  execute these targets in the next order first:
> 
> 
> 
>   • ant generate-cql-html
>   • ant maven-ant-tasks-init
> and then  execute ant default target
> 
> 
>   • ant
> 
> 
> the project build succesfully.
> 
> Regards.
> 
> 
>  
> 
> 
> note:
> 
> I do not have this issue in my mac.
> 
> 
> 
> 
> 
> 
> Miguel Angel Martín Junquera
> Analyst Engineer.
> miguelangel.mar...@brainsins.com
> 



Re: Connecting to a remote cassandra node..

2013-09-30 Thread Krishna Chaitanya
Thanks...
On Sep 30, 2013 12:35 PM, "Aaron Morton"  wrote:

> See the comments for rpc_address and listen_address in the yaml file.
> These control which interfaces the server binds to.
>
> If you leave them as blank and DNS is correctly setup they will bind to
> the correct address.
>
> If you set them to 0.0.0.0 they will bind to all interfaces, you should
> not do this in production.
>
> Cheers
>
> -
> Aaron Morton
> New Zealand
> @aaronmorton
>
> Co-Founder & Principal Consultant
> Apache Cassandra Consulting
> http://www.thelastpickle.com
>
> On 27/09/2013, at 8:27 PM, Krishna Chaitanya 
> wrote:
>
> > Thank you...but cant we dynamically connect to any system instead of
> changing the .yaml file each time?? I want to take the input from the user
> regarding the host and port that he wants to use...Thanks.
> >
> > On Sep 27, 2013 1:38 PM, "Kais Ahmed"  wrote:
> > hello,
> >
> > you have to check listen_address in cassandra.yaml, change the localhost
> value by the ip of the machine and restart cassandra
> >
> >
> > 2013/9/27 Krishna Chaitanya 
> > Hello,
> >   I am relatively new to cassandra. I am using a library called
> libQtCassandra for accesing the cassandra database from my c++ programs.
> When I try to connect to the localhost cassandra , everything seems fine
> but when I try to connect to a remote node on which cassandra is up and
> running , it says connection refused. Any help would be of a great value.
> Thank You...
> >
> > --
> > Regards,
> > BNSK.
> >
>
>


Re: What is the best way to install & upgrade Cassandra on Ubuntu ?

2013-09-30 Thread Ertio Lew
Thanks Aaron!

Does DSC include other things like Opscenter by default ? I installed DSC
on linux, but Opscenter wasn't installed there but when tried on Windows it
was installed along with JRE & python, using the windows installer.

Would it be possible to remove any of these installations but keeping the
data intact & easily switch to the another, I mean switching from DSC
package to apache one or vice versa ?


On Mon, Sep 30, 2013 at 1:10 PM, Aaron Morton wrote:

> I am not sure if I should use datastax's DSC or official Debian packages
> from Cassandra. How do I choose between them for a production server ?
>
> They are technically the same.
> The DSC update will come out a little after the Apache release, and I
> _think_ they release for every Apache release.
>
>  1.  when I upgrade to a newer version, would that retain my previous
> configurations so that I don't need to configure everything again ?
>
> Yes if you select that when doing the package install.
>
> 2.  would that smoothly replace the previous installation by itself ?
>>
>
>> Yes
>>
>
> 3.  what's the way (kindly, if you can tell the command) to upgrade ?
>>
>
>>
> http://www.datastax.com/documentation/cassandra/2.0/webhelp/index.html#upgrade/upgradeC_c.html#concept_ds_yqj_5xr_ck
>
> 4. when should I prefer datastax's dsc to that ? (I need to install for
>> production env.)
>>
> Above
>
> Hope that helps.
>
>
> -
> Aaron Morton
> New Zealand
> @aaronmorton
>
> Co-Founder & Principal Consultant
> Apache Cassandra Consulting
> http://www.thelastpickle.com
>
> On 27/09/2013, at 11:01 PM, Ertio Lew  wrote:
>
> I am not sure if I should use datastax's DSC or official Debian packages
> from Cassandra. How do I choose between them for a production server ?
>
>
>
> On Fri, Sep 27, 2013 at 11:02 AM, Ertio Lew  wrote:
>
>>
>>  Could you please clarify that:
>> 1.  when I upgrade to a newer version, would that retain my previous
>> configurations so that I don't need to configure everything again ?
>> 2.  would that smoothly replace the previous installation by itself ?
>> 3.  what's the way (kindly, if you can tell the command) to upgrade ?
>> 4. when should I prefer datastax's dsc to that ? (I need to install for
>> production env.)
>>
>>
>> On Fri, Sep 27, 2013 at 12:50 AM, Robert Coli wrote:
>>
>>> On Thu, Sep 26, 2013 at 12:05 PM, Ertio Lew  wrote:
>>>
 How do you install Cassandra on Ubuntu & later how do you upgrade the
 installation on the node when an update has arrived ? Do you simply
 download & replace the latest tar.gz, untar it to replace the older
 cassandra files? How do you do it ? How does this upgrade process differ
 for a major version upgrade, like say switching from 1.2 series to 2.0
 series ?

>>>
>>> Use the deb packages. To upgrade, install the new package. Only upgrade
>>> a single major version. and be sure to consult NEWS.txt for any upgrade
>>> caveats.
>>>
>>> Also be aware of this sub-optimal behavior of the debian packages :
>>>
>>> https://issues.apache.org/jira/browse/CASSANDRA-2356
>>>
>>> =Rob
>>>
>>>
>>
>
>


RE: 2.0.1 counter replicate on write error

2013-09-30 Thread Christopher Wirt
Thanks Aaron, I've added to the ticket. We were not running on TRACE
logging. 

 

From: Aaron Morton [mailto:aa...@thelastpickle.com] 
Sent: 30 September 2013 08:37
To: user@cassandra.apache.org
Subject: Re: 2.0.1 counter replicate on write error

 

ERROR [ReplicateOnWriteStage:19] 2013-09-27 10:17:14,778
CassandraDaemon.java (line 185) Exception in thread
Thread[ReplicateOnWriteStage:19,5,main]

java.lang.AssertionError: DecoratedKey(-1754949563326053382,
a414b0c07f0547f8a75410555716ced6) != DecoratedKey(-1754949563326053382,
aeadcec8184445d4ab631ef4250927d0) in
/disk3/cassandra/data/struqrealtime/counters/struqrealtime-counters-jb-83195
3-Data.db

at
org.apache.cassandra.db.columniterator.SSTableNamesIterator.read(SSTableName
sIterator.java:114)

at
org.apache.cassandra.db.columniterator.SSTableNamesIterator.(SSTableNa
mesIterator.java:62)

 

When reading from an SSTable the position returned from the -Index.db / KEYS
cache pointed to a row in the -Data.db component that was for a different
row. 

 

DecoratedKey(-1754949563326053382, aeadcec8184445d4ab631ef4250927d0)

Was what we were searching for

 

DecoratedKey(-1754949563326053382, a414b0c07f0547f8a75410555716ced6)

Is what was found in the data component. 

 

The first part is the Token (M3 hash) the second is the key. It looks like a
collision, but it could also be a bug somewhere else. 

 

Code in SSTableReader.getPosition() points to
https://issues.apache.org/jira/browse/CASSANDRA-4687 and adds an assertion
that is only trigger if TRACE logging is running. Can you add to the 4687
ticket and update the thread ? 

 

Cheers

 

-

Aaron Morton

New Zealand

@aaronmorton

 

Co-Founder & Principal Consultant

Apache Cassandra Consulting

http://www.thelastpickle.com

 

On 27/09/2013, at 10:50 PM, Christopher Wirt  wrote:





Hello,

 

I've started to see a slightly worrying error appear in our logs
occasionally. We're writing at 400qps per machine and I only see this appear
every 5-10minutes.

 

Seems to have started when I switched us to using the hsha thrift server
this morning. We've been running 2.0.1 ran off the sync thrift server since
yesterday without seeing this error.  But might not be related.

 

There are some machines in another DC still running 1.2.10.

 

Anyone seen this before? Have any insight?

 

ERROR [ReplicateOnWriteStage:19] 2013-09-27 10:17:14,778
CassandraDaemon.java (line 185) Exception in thread
Thread[ReplicateOnWriteStage:19,5,main]

java.lang.AssertionError: DecoratedKey(-1754949563326053382,
a414b0c07f0547f8a75410555716ced6) != DecoratedKey(-1754949563326053382,
aeadcec8184445d4ab631ef4250927d0) in
/disk3/cassandra/data/struqrealtime/counters/struqrealtime-counters-jb-83195
3-Data.db

at
org.apache.cassandra.db.columniterator.SSTableNamesIterator.read(SSTableName
sIterator.java:114)

at
org.apache.cassandra.db.columniterator.SSTableNamesIterator.(SSTableNa
mesIterator.java:62)

at
org.apache.cassandra.db.filter.NamesQueryFilter.getSSTableColumnIterator(Nam
esQueryFilter.java:87)

at
org.apache.cassandra.db.filter.QueryFilter.getSSTableColumnIterator(QueryFil
ter.java:62)

at
org.apache.cassandra.db.CollationController.collectAllData(CollationControll
er.java:249)

at
org.apache.cassandra.db.CollationController.getTopLevelColumns(CollationCont
roller.java:53)

at
org.apache.cassandra.db.ColumnFamilyStore.getTopLevelColumns(ColumnFamilySto
re.java:1468)

at
org.apache.cassandra.db.ColumnFamilyStore.getColumnFamily(ColumnFamilyStore.
java:1294)

at org.apache.cassandra.db.Keyspace.getRow(Keyspace.java:332)

at
org.apache.cassandra.db.SliceByNamesReadCommand.getRow(SliceByNamesReadComma
nd.java:55)

at
org.apache.cassandra.db.CounterMutation.makeReplicationMutation(CounterMutat
ion.java:100)

at
org.apache.cassandra.service.StorageProxy$8$1.runMayThrow(StorageProxy.java:
1107)

at
org.apache.cassandra.service.StorageProxy$DroppableRunnable.run(StorageProxy
.java:1897)

at
java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:11
45)

at
java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:6
15)

at java.lang.Thread.run(Thread.java:724)

 



Re: temporal solution to CASSANDRA-5543: BUILD FAILED at gen-cql2-grammar target

2013-09-30 Thread Miguel Angel Martin junquera
hi:



is that mean that antlr-3.2.jar is not the correct version?

what is the correct version?

In the build file,  I see that cassandra uses the jar lib at the *
${build.lib}*  folder, in this case antlr-3.2.jar


...

 
  Building Grammar
${build.src.java}/org/apache/cassandra/cql/Cql.g  ...
  
 
 
 
  


...



thanks in advance


Miguel Angel Martín Junquera
Analyst Engineer.
miguelangel.mar...@brainsins.com



2013/9/30 Aaron Morton 

> It's an error in the antlr compilation, check the antlr versions.
>
> Cheers
>
> -
> Aaron Morton
> New Zealand
> @aaronmorton
>
> Co-Founder & Principal Consultant
> Apache Cassandra Consulting
> http://www.thelastpickle.com
>
> On 27/09/2013, at 11:53 PM, Miguel Angel Martin junquera <
> mianmarjun.mailingl...@gmail.com> wrote:
>
> >
> >
> >
> >
> > hi all:
> >
> > Environment
> >
> >   • apache-cassandra-2.0.1-src
> >   • EC2
> >   • Linux version 3.2.30-49.59.amzn1.x86_64
> (mockbuild@gobi-build-31003) (gcc version 4.4.6 20110731 (Red Hat 4.4.6-3)
> >
> > When i try to build apache-cassandra-2.0.1-src in EC2 red had AMI, I
>  have this error at  the target, gen-cql2-grammar:
> >
> >
> > gen-cql2-grammar:
> >  [echo] Building Grammar
> /home/ec2-user/apache-cassandra/src/java/org/apache/cassandra/cql/Cql.g
> >
> > ….
> >
> >   [java] warning(209):
> /home/ec2-user/apache-cassandra/src/java/org/apache/cassandra/cql/Cql.g:638:1:
> Multiple token rules can match input such as "'0'..'9'": INTEGER, FLOAT,
> UUID
> >  [java]
> >  [java] As a result, token(s) FLOAT,UUID were disabled for that input
> >  [java] warning(209):
> /home/ec2-user/apache-cassandra/src/java/org/apache/cassandra/cql/Cql.g:634:1:
> Multiple token rules can match input such as "'I'": K_INSERT, K_IN,
> K_INDEX, K_INTO, IDENT, COMPIDENT
> >  [java]
> >  [java] As a result, token(s) K_IN,K_INDEX,K_INTO,IDENT,COMPIDENT
> were disabled for that input
> >  [java] warning(209):
> /home/ec2-user/apache-cassandra/src/java/org/apache/cassandra/cql/Cql.g:634:1:
> Multiple token rules can match input such as "{'R', 'r'}": K_REVERSED,
> IDENT, COMPIDENT
> >  [java]
> >  [java] As a result, token(s) IDENT,COMPIDENT were disabled for that
> input
> >  [java] warning(209):
> /home/ec2-user/apache-cassandra/src/java/org/apache/cassandra/cql/Cql.g:634:1:
> Multiple token rules can match input such as "'T'": K_LEVEL, K_TRUNCATE,
> K_COLUMNFAMILY, K_TIMESTAMP, K_TTL, K_TYPE, IDENT, COMPIDENT
> >  [java]
> >  [java] As a result, token(s)
> K_TRUNCATE,K_COLUMNFAMILY,K_TIMESTAMP,K_TTL,K_TYPE,IDENT,COMPIDENT were
> disabled for that input
> >  [java] error(208):
> /home/ec2-user/apache-cassandra/src/java/org/apache/cassandra/cql/Cql.g:654:1:
> The following token definitions can never be matched because prior tokens
> match the same input:
> T__93,T__94,T__97,T__98,T__101,T__105,T__107,K_WITH,K_USING,K_USE,K_FIRST,K_COUNT,K_SET,K_APPLY,K_BATCH,K_TRUNCATE,K_IN,K_CREATE,K_KEYSPACE,K_COLUMNFAMILY,K_INDEX,K_ON,K_DROP,K_INTO,K_TIMESTAMP,K_TTL,K_ALTER,K_ADD,K_TYPE,RANGEOP,FLOAT,COMPIDENT,UUID,MULTILINE_COMMENT
> >
> > BUILD FAILED
> > /home/ec2-user/apache-cassandra/build.xml:218: Java returned: 1
> >
> >
> >
> >
> >
> >
> > If  I  execute these targets in the next order first:
> >
> >
> >
> >   • ant generate-cql-html
> >   • ant maven-ant-tasks-init
> > and then  execute ant default target
> >
> >
> >   • ant
> >
> >
> > the project build succesfully.
> >
> > Regards.
> >
> >
> >
> >
> >
> > note:
> >
> > I do not have this issue in my mac.
> >
> >
> >
> >
> >
> >
> > Miguel Angel Martín Junquera
> > Analyst Engineer.
> > miguelangel.mar...@brainsins.com
> >
>
>


Re: Among Datastax community & Cassandra debian package, which to choose for production install ?

2013-09-30 Thread Ken Hancock
OpsCenter should be a separate package as you would only install it on a
single node, not necessarily even one that is running Cassandra.




On Sat, Sep 28, 2013 at 2:12 PM, Ertio Lew  wrote:

> I think both provide the same thing except Datastax Community also
> provides some extras like Opscenter, etc. But I cannot find opscenter
> installed when I installled DSC on ubuntu. Although on windows
> installation, I saw opscenter & JRE as well , so I think for DSC, there is
> no such prerequisite for Oracle JRE as required for Cassandra debain
> package, is it so ?
>
> Btw which is usually preferred for production installs ?
>
> I may need to use Opscenter but just *occasionally*.
>



-- 
*Ken Hancock *| System Architect, Advanced Advertising
SeaChange International
50 Nagog Park
Acton, Massachusetts 01720
ken.hanc...@schange.com | www.schange.com |
NASDAQ:SEAC

Office: +1 (978) 889-3329 | [image: Google Talk:]
ken.hanc...@schange.com | [image:
Skype:]hancockks | [image: Yahoo IM:]hancockks [image:
LinkedIn]

[image: SeaChange International]
 This e-mail and any attachments may contain
information which is SeaChange International confidential. The information
enclosed is intended only for the addressees herein and may not be copied
or forwarded without permission from SeaChange International.


Re: Among Datastax community & Cassandra debian package, which to choose for production install ?

2013-09-30 Thread Ertio Lew
& what about JRE is it provided by DSC so that I don't need to take care of
those* *Oracle JRE* updates myself ? So which one is more preferable or
lets say commonly used for production installs ?

Btw I think I should be able to easily switch between them retaining data ?


On Mon, Sep 30, 2013 at 6:04 PM, Ken Hancock wrote:

> OpsCenter should be a separate package as you would only install it on a
> single node, not necessarily even one that is running Cassandra.
>
>
>
>
> On Sat, Sep 28, 2013 at 2:12 PM, Ertio Lew  wrote:
>
>> I think both provide the same thing except Datastax Community also
>> provides some extras like Opscenter, etc. But I cannot find opscenter
>> installed when I installled DSC on ubuntu. Although on windows
>> installation, I saw opscenter & JRE as well , so I think for DSC, there is
>> no such prerequisite for Oracle JRE as required for Cassandra debain
>> package, is it so ?
>>
>> Btw which is usually preferred for production installs ?
>>
>> I may need to use Opscenter but just *occasionally*.
>>
>
>
>
> --
> *Ken Hancock *| System Architect, Advanced Advertising
> SeaChange International
> 50 Nagog Park
> Acton, Massachusetts 01720
> ken.hanc...@schange.com | www.schange.com | 
> NASDAQ:SEAC
>
> Office: +1 (978) 889-3329 | [image: Google Talk:] ken.hanc...@schange.com
>  | [image: Skype:]hancockks | [image: Yahoo IM:]hancockks [image:
> LinkedIn] 
>
> [image: SeaChange International]
>  This e-mail and any attachments may contain
> information which is SeaChange International confidential. The information
> enclosed is intended only for the addressees herein and may not be copied
> or forwarded without permission from SeaChange International.
>


Cassandra Summit EU 2013

2013-09-30 Thread Jeremy Hanna
For those in the Europe area, there will be a Cassandra Summit EU 2013 in 
London in the month of October.  On 17 October, there will be the main 
conference sessions and the 16th and 18th there will be Cassandra workshops.

http://www.datastax.com/cassandraeurope2013

The speakers have been announced and the presentation abstracts are all on 
there.  Like always, the presentations will be recorded and posted on Planet 
Cassandra, but it's great to meet and interact with people in the community - 
in my opinion that's the best part of any conference.

Anyway, just wanted to make sure people knew.

Cheers,

Jeremy

Undefined name key_alias in selection clause pig 0.11.1 and cassandra 2.0.1

2013-09-30 Thread Miguel Angel Martin junquera
hi:


I try to test cassandra 2.0.1 and pig 0.11.1
but always when i try to load  any keyspacei have this error:

Example creating a test keyspace and data:



I have this error:




*grunt> define CqlStorage org.apache.cassandra.hadoop.pig.CqlStorage();*
*grunt> rows = LOAD 'cql://keyspace0/test' USING
org.apache.cassandra.hadoop.pig.CqlStorage();*
*2013-09-30 15:43:09,176 [main] ERROR org.apache.pig.tools.grunt.Grunt -
ERROR 2999: Unexpected internal error.
InvalidRequestException(why:Undefined name key_alias in selection clause)*
*Details at logfile: /private/var/log/pig/logging_pig_0.11.1.log*




and in the log file there are not  more info



Pig Stack Trace
---
ERROR 2999: Unexpected internal error.
InvalidRequestException(why:Undefined name key_alias in selection clause)

java.lang.RuntimeException: InvalidRequestException(why:Undefined name
key_alias in selection clause)
at
org.apache.cassandra.hadoop.pig.AbstractCassandraStorage.initSchema(AbstractCassandraStorage.java:511)
at
org.apache.cassandra.hadoop.pig.CqlStorage.setLocation(CqlStorage.java:246)
at
org.apache.cassandra.hadoop.pig.CqlStorage.getSchema(CqlStorage.java:280)
at
org.apache.pig.newplan.logical.relational.LOLoad.getSchemaFromMetaData(LOLoad.java:151)
at
org.apache.pig.newplan.logical.relational.LOLoad.getSchema(LOLoad.java:110)
at
org.apache.pig.newplan.logical.visitor.LineageFindRelVisitor.visit(LineageFindRelVisitor.java:100)
at
org.apache.pig.newplan.logical.relational.LOLoad.accept(LOLoad.java:219)
at
org.apache.pig.newplan.DependencyOrderWalker.walk(DependencyOrderWalker.java:75)
at org.apache.pig.newplan.PlanVisitor.visit(PlanVisitor.java:50)
at
org.apache.pig.newplan.logical.visitor.CastLineageSetter.(CastLineageSetter.java:57)
at org.apache.pig.PigServer$Graph.compile(PigServer.java:1635)
at org.apache.pig.PigServer$Graph.validateQuery(PigServer.java:1566)
at org.apache.pig.PigServer$Graph.registerQuery(PigServer.java:1538)
at org.apache.pig.PigServer.registerQuery(PigServer.java:540)
at
org.apache.pig.tools.grunt.GruntParser.processPig(GruntParser.java:970)
at
org.apache.pig.tools.pigscript.parser.PigScriptParser.parse(PigScriptParser.java:386)
at
org.apache.pig.tools.grunt.GruntParser.parseStopOnError(GruntParser.java:189)
at
org.apache.pig.tools.grunt.GruntParser.parseStopOnError(GruntParser.java:165)
at org.apache.pig.tools.grunt.Grunt.run(Grunt.java:69)
at org.apache.pig.Main.run(Main.java:490)
at org.apache.pig.Main.main(Main.java:111)
at sun.reflect.NativeMethodAccessorImpl.invoke0(Native Method)
at
sun.reflect.NativeMethodAccessorImpl.invoke(NativeMethodAccessorImpl.java:57)
at
sun.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:43)
at java.lang.reflect.Method.invoke(Method.java:606)
at org.apache.hadoop.util.RunJar.main(RunJar.java:156)
Caused by: InvalidRequestException(why:Undefined name key_alias in
selection clause)
at
org.apache.cassandra.thrift.Cassandra$execute_cql3_query_result$execute_cql3_query_resultStandardScheme.read(Cassandra.java:48006)
at
org.apache.cassandra.thrift.Cassandra$execute_cql3_query_result$execute_cql3_query_resultStandardScheme.read(Cassandra.java:47983)
at
org.apache.cassandra.thrift.Cassandra$execute_cql3_query_result.read(Cassandra.java:47898)
at
org.apache.thrift.TServiceClient.receiveBase(TServiceClient.java:78)
at
org.apache.cassandra.thrift.Cassandra$Client.recv_execute_cql3_query(Cassandra.java:1658)
at
org.apache.cassandra.thrift.Cassandra$Client.execute_cql3_query(Cassandra.java:1643)
at
org.apache.cassandra.hadoop.pig.AbstractCassandraStorage.getCfDef(AbstractCassandraStorage.java:573)
at
org.apache.cassandra.hadoop.pig.AbstractCassandraStorage.initSchema(AbstractCassandraStorage.java:500)
... 25 more




any idea?


thanks




Miguel Angel Martín Junquera
Analyst Engineer.
miguelangel.mar...@brainsins.com


Re: Undefined name key_alias in selection clause pig 0.11.1 and cassandra 2.0.1

2013-09-30 Thread Miguel Angel Martin junquera
I have the same issue using cassandraStorage()  and other dataset for this


*note*:

the example cql3 datatest are



CREATE TABLE test (
  id text PRIMARY KEY,
  title text,
  age int
);
and insert some dummy data

insert into test (id, title, age) values('1', 'child', 21);
insert into test (id, title, age) values('2', 'support', 21);
insert into test (id, title, age) values('3', 'manager', 31);
insert into test (id, title, age) values('4', 'QA', 41);
insert into test (id, title, age) values('5', 'QA', 30);
insert into test (id, title, age) values('6', 'QA', 30);




Miguel Angel Martín Junquera
Analyst Engineer.
miguelangel.mar...@brainsins.com



2013/9/30 Miguel Angel Martin junquera 

>
>
>
> hi:
>
>
> I try to test cassandra 2.0.1 and pig 0.11.1
> but always when i try to load  any keyspacei have this error:
>
> Example creating a test keyspace and data:
>
>
>
> I have this error:
>
>
>
>
> *grunt> define CqlStorage org.apache.cassandra.hadoop.pig.CqlStorage();*
> *grunt> rows = LOAD 'cql://keyspace0/test' USING
> org.apache.cassandra.hadoop.pig.CqlStorage();*
> *2013-09-30 15:43:09,176 [main] ERROR org.apache.pig.tools.grunt.Grunt -
> ERROR 2999: Unexpected internal error.
> InvalidRequestException(why:Undefined name key_alias in selection clause)*
> *Details at logfile: /private/var/log/pig/logging_pig_0.11.1.log*
>
>
>
>
> and in the log file there are not  more info
>
>
>
> Pig Stack Trace
> ---
> ERROR 2999: Unexpected internal error.
> InvalidRequestException(why:Undefined name key_alias in selection clause)
>
> java.lang.RuntimeException: InvalidRequestException(why:Undefined name
> key_alias in selection clause)
> at
> org.apache.cassandra.hadoop.pig.AbstractCassandraStorage.initSchema(AbstractCassandraStorage.java:511)
> at
> org.apache.cassandra.hadoop.pig.CqlStorage.setLocation(CqlStorage.java:246)
> at
> org.apache.cassandra.hadoop.pig.CqlStorage.getSchema(CqlStorage.java:280)
> at
> org.apache.pig.newplan.logical.relational.LOLoad.getSchemaFromMetaData(LOLoad.java:151)
> at
> org.apache.pig.newplan.logical.relational.LOLoad.getSchema(LOLoad.java:110)
> at
> org.apache.pig.newplan.logical.visitor.LineageFindRelVisitor.visit(LineageFindRelVisitor.java:100)
> at
> org.apache.pig.newplan.logical.relational.LOLoad.accept(LOLoad.java:219)
> at
> org.apache.pig.newplan.DependencyOrderWalker.walk(DependencyOrderWalker.java:75)
> at org.apache.pig.newplan.PlanVisitor.visit(PlanVisitor.java:50)
> at
> org.apache.pig.newplan.logical.visitor.CastLineageSetter.(CastLineageSetter.java:57)
> at org.apache.pig.PigServer$Graph.compile(PigServer.java:1635)
> at
> org.apache.pig.PigServer$Graph.validateQuery(PigServer.java:1566)
> at
> org.apache.pig.PigServer$Graph.registerQuery(PigServer.java:1538)
> at org.apache.pig.PigServer.registerQuery(PigServer.java:540)
> at
> org.apache.pig.tools.grunt.GruntParser.processPig(GruntParser.java:970)
> at
> org.apache.pig.tools.pigscript.parser.PigScriptParser.parse(PigScriptParser.java:386)
> at
> org.apache.pig.tools.grunt.GruntParser.parseStopOnError(GruntParser.java:189)
> at
> org.apache.pig.tools.grunt.GruntParser.parseStopOnError(GruntParser.java:165)
> at org.apache.pig.tools.grunt.Grunt.run(Grunt.java:69)
> at org.apache.pig.Main.run(Main.java:490)
> at org.apache.pig.Main.main(Main.java:111)
> at sun.reflect.NativeMethodAccessorImpl.invoke0(Native Method)
> at
> sun.reflect.NativeMethodAccessorImpl.invoke(NativeMethodAccessorImpl.java:57)
> at
> sun.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:43)
> at java.lang.reflect.Method.invoke(Method.java:606)
> at org.apache.hadoop.util.RunJar.main(RunJar.java:156)
> Caused by: InvalidRequestException(why:Undefined name key_alias in
> selection clause)
> at
> org.apache.cassandra.thrift.Cassandra$execute_cql3_query_result$execute_cql3_query_resultStandardScheme.read(Cassandra.java:48006)
> at
> org.apache.cassandra.thrift.Cassandra$execute_cql3_query_result$execute_cql3_query_resultStandardScheme.read(Cassandra.java:47983)
> at
> org.apache.cassandra.thrift.Cassandra$execute_cql3_query_result.read(Cassandra.java:47898)
> at
> org.apache.thrift.TServiceClient.receiveBase(TServiceClient.java:78)
> at
> org.apache.cassandra.thrift.Cassandra$Client.recv_execute_cql3_query(Cassandra.java:1658)
> at
> org.apache.cassandra.thrift.Cassandra$Client.execute_cql3_query(Cassandra.java:1643)
> at
> org.apache.cassandra.hadoop.pig.AbstractCassandraStorage.getCfDef(AbstractCassandraStorage.java:573)
> at
> org.apache.cassandra.hadoop.pig.AbstractCassandraStorage.initSchema(AbstractCassandraStorage.java:500)
> ... 25 more
>
>
>
>
> any idea?
>
>
> thanks
>
>
>
>
> Miguel Angel Martín Junquera
> Analyst Engineer

Why Solandra stores Solr data in Cassandra ? Isn't solr complete solution ?

2013-09-30 Thread Ertio Lew
Solr's data is stored on the file system as a set of index files[
http://stackoverflow.com/a/7685579/530153]. Then why do we need anything
like Solandra or DataStax Enterprise Search? Isn't Solr complete solution
in itself ?  What do we need to integrate with Cassandra ?


Re: Why Solandra stores Solr data in Cassandra ? Isn't solr complete solution ?

2013-09-30 Thread Peter Lin
the main reason is scalability and performance.

If your Solr indexes fit fine on a single system and doesn't need to scale
out, Cassandra/HDFS isn't necessary.


On Mon, Sep 30, 2013 at 11:50 AM, Ertio Lew  wrote:

> Solr's data is stored on the file system as a set of index files[
> http://stackoverflow.com/a/7685579/530153]. Then why do we need anything
> like Solandra or DataStax Enterprise Search? Isn't Solr complete solution
> in itself ?  What do we need to integrate with Cassandra ?
>
>
>
>


Re: Why Solandra stores Solr data in Cassandra ? Isn't solr complete solution ?

2013-09-30 Thread Robert Coli
On Mon, Sep 30, 2013 at 8:50 AM, Ertio Lew  wrote:

> Solr's data is stored on the file system as a set of index files[
> http://stackoverflow.com/a/7685579/530153]. Then why do we need anything
> like Solandra or DataStax Enterprise Search? Isn't Solr complete solution
> in itself ?  What do we need to integrate with Cassandra ?
>

Solr's index sitting on a single machine, even if that single machine can
vertically scale, is a single point of failure.

The value add of DES is that the index has the same availability
characteristics as the underlying data, because it is stored in the same
cluster.

=Rob


Re: Why Solandra stores Solr data in Cassandra ? Isn't solr complete solution ?

2013-09-30 Thread Ken Hancock
To clarify, solr indexes are not distributed in the same way that Cassandra
data is stored.

With Cassandra, each node receives a fraction of the keyspace (based on
your replication factor and token assignment).  With DSE Search, writes to
Cassandra are hooked and each node independently indexes its data and keeps
this index on the local file system.  If you have keyspace with RF=3 then
three nodes will index each document. Indexes, unlike from Solr, only store
the docids and the actual field values are stored in Cassandra.

When it comes to search, DSE splits up the search so in the example above
only one of those RF=3 nodes will be queried for a particular token range
so that data can be unioned across all the nodes with different token
ranges.

Not sure about Solandra, but you do need to be aware that there's a number
of Solr search options that are not supported on distribute searches/DSE
Search.

http://wiki.apache.org/solr/DistributedSearch
http://wiki.apache.org/solr/FieldCollapsing

Also, be aware that while Cassandra has knobs to allow you to get
consistent read results (CL=QUORUM), DSE Search does not. If a node drops
messages for whatever reason, outtage, mutation, etc. its solr indexes will
be inconsistent with other nodes in its replication group.



On Mon, Sep 30, 2013 at 1:06 PM, Robert Coli  wrote:

> On Mon, Sep 30, 2013 at 8:50 AM, Ertio Lew  wrote:
>
>> Solr's data is stored on the file system as a set of index files[
>> http://stackoverflow.com/a/7685579/530153]. Then why do we need anything
>> like Solandra or DataStax Enterprise Search? Isn't Solr complete solution
>> in itself ?  What do we need to integrate with Cassandra ?
>>
>
> Solr's index sitting on a single machine, even if that single machine can
> vertically scale, is a single point of failure.
>
> The value add of DES is that the index has the same availability
> characteristics as the underlying data, because it is stored in the same
> cluster.
>
> =Rob
>
>



-- 
*Ken Hancock *| System Architect, Advanced Advertising
SeaChange International
50 Nagog Park
Acton, Massachusetts 01720
ken.hanc...@schange.com | www.schange.com |
NASDAQ:SEAC

Office: +1 (978) 889-3329 | [image: Google Talk:]
ken.hanc...@schange.com | [image:
Skype:]hancockks | [image: Yahoo IM:]hancockks [image:
LinkedIn]

[image: SeaChange International]
 This e-mail and any attachments may contain
information which is SeaChange International confidential. The information
enclosed is intended only for the addressees herein and may not be copied
or forwarded without permission from SeaChange International.


Re: Why Solandra stores Solr data in Cassandra ? Isn't solr complete solution ?

2013-09-30 Thread Andrey Ilinykh
> Also, be aware that while Cassandra has knobs to allow you to get
> consistent read results (CL=QUORUM), DSE Search does not. If a node drops
> messages for whatever reason, outtage, mutation, etc. its solr indexes will
> be inconsistent with other nodes in its replication group.
>
> Will repair fix it?


Re: Why Solandra stores Solr data in Cassandra ? Isn't solr complete solution ?

2013-09-30 Thread Ken Hancock
Yes.


On Mon, Sep 30, 2013 at 1:57 PM, Andrey Ilinykh  wrote:

>
> Also, be aware that while Cassandra has knobs to allow you to get
>> consistent read results (CL=QUORUM), DSE Search does not. If a node drops
>> messages for whatever reason, outtage, mutation, etc. its solr indexes will
>> be inconsistent with other nodes in its replication group.
>>
>> Will repair fix it?
>



-- 
*Ken Hancock *| System Architect, Advanced Advertising
SeaChange International
50 Nagog Park
Acton, Massachusetts 01720
ken.hanc...@schange.com | www.schange.com |
NASDAQ:SEAC

Office: +1 (978) 889-3329 | [image: Google Talk:]
ken.hanc...@schange.com | [image:
Skype:]hancockks | [image: Yahoo IM:]hancockks [image:
LinkedIn]

[image: SeaChange International]
 This e-mail and any attachments may contain
information which is SeaChange International confidential. The information
enclosed is intended only for the addressees herein and may not be copied
or forwarded without permission from SeaChange International.


How to log the details of the updated data locally

2013-09-30 Thread sathiya prabhu
Hi all,

In cassandra cluster, once a write/update is successful to the particular 
instance locally, i want to log the data that is updated and its timestamps 
separately in a file.. In which class it will be more appropriate to do this.. 

To the extent i explored the codebase, it's possible to do that in Keyspace 
class (apply method) in db package.. But i don't know how to retrieve the 
timestamp details from the mutation object.. Is Keyspace class is appropriate 
for my purpose.. If yes, please kindly provide me some ideas to retrieve 
timestamp details from mutation object(RowMutation)..

Any help is appreciable.. Looking forward for your kind replies.. Thanks in 
advance..

Thank you.

Re: Why Solandra stores Solr data in Cassandra ? Isn't solr complete solution ?

2013-09-30 Thread Sávio Teles
> Solr's index sitting on a single machine, even if that single machine can
> vertically scale, is a single point of failure.
>

And about Cloud Solr?


2013/9/30 Ken Hancock 

> Yes.
>
>
> On Mon, Sep 30, 2013 at 1:57 PM, Andrey Ilinykh wrote:
>
>>
>> Also, be aware that while Cassandra has knobs to allow you to get
>>> consistent read results (CL=QUORUM), DSE Search does not. If a node drops
>>> messages for whatever reason, outtage, mutation, etc. its solr indexes will
>>> be inconsistent with other nodes in its replication group.
>>>
>>> Will repair fix it?
>>
>
>
>
> --
> *Ken Hancock *| System Architect, Advanced Advertising
> SeaChange International
> 50 Nagog Park
> Acton, Massachusetts 01720
> ken.hanc...@schange.com | www.schange.com | 
> NASDAQ:SEAC
>
> Office: +1 (978) 889-3329 | [image: Google Talk:] ken.hanc...@schange.com
>  | [image: Skype:]hancockks | [image: Yahoo IM:]hancockks [image:
> LinkedIn] 
>
> [image: SeaChange International]
>  This e-mail and any attachments may contain
> information which is SeaChange International confidential. The information
> enclosed is intended only for the addressees herein and may not be copied
> or forwarded without permission from SeaChange International.
>



-- 
Atenciosamente,
Sávio S. Teles de Oliveira
voice: +55 62 9136 6996
http://br.linkedin.com/in/savioteles
Mestrando em Ciências da Computação - UFG
Arquiteto de Software
Laboratory for Ubiquitous and Pervasive Applications (LUPA) - UFG


Segmentation fault when trying to store into cassandra...

2013-09-30 Thread Krishna Chaitanya
Hello,
   I modified a network probe which collects network packets to
store them into cassandra. So there are many packets that are coming in, I
capture the packets in the program and store them into cassandra. I am
using libQtCassandra library. The program is crashing with segmentation
fault as soon as I run it. Can someone help as to what all can go wrong
here?? Could there be a problem with row/col keys or is it some
configuration parameter or the speed at which the packets or coming? I am
not able to figure it out. Thank you.

-- 
Regards,
BNSK*.
*


Re: Segmentation fault when trying to store into cassandra...

2013-09-30 Thread Vivek Mishra
Java version issue?
Using sun jdk or open jdk?

-Vivek


On Tue, Oct 1, 2013 at 6:16 AM, Krishna Chaitanya wrote:

> Hello,
>I modified a network probe which collects network packets to
> store them into cassandra. So there are many packets that are coming in, I
> capture the packets in the program and store them into cassandra. I am
> using libQtCassandra library. The program is crashing with segmentation
> fault as soon as I run it. Can someone help as to what all can go wrong
> here?? Could there be a problem with row/col keys or is it some
> configuration parameter or the speed at which the packets or coming? I am
> not able to figure it out. Thank you.
>
> --
> Regards,
> BNSK*.
> *
>


Why Cassandra so depend on client local timestamp?

2013-09-30 Thread Jason Tang
Following case may be logical correct for Cassandra, but difficult for user.
Let's say:

Cassandra consistency level: write all, read one
replication_factor:3

For one record, rowkey:001, column:status

Client 1, insert value for rowkey 001, status:True, timestamp 11:00:05
Client 2 Slice Query, get the value True for rowkey 001, @11:00:00
Client 2, update value for rowkey 001, status:False, timestamp 11:00:02

So the client update sequence is True to False, although the update
requests are from different nodes, but the sequence are logically ordered.

But the result is rowkey:001, column:status, value: True

So why Cassandra so depend on client local time? Why not using server
localtime instead client local time?

Because I am using consistency level write all, and replication_factor:3,
so for all the 3 nodes, the update sequence is correct (True -> False),
they can give a correct final results.

If for some reason, it need strong depends on operation's timestamp, then
query operation also need a timestamp, then Client 2 will not see the value
True, which happen in "future".

So either using server timestamp or provide a consistent view by using
timestamp for query, it will be more consistent.

Otherwise, the consistency of Cassandra is so weak.