Re: Snapshot failing on JSON files in 1.1.0
I have the same error with the last Datastax AMI (1.0.9). Is that the same bug ? Requested snapshot for: cassa_teads Exception in thread "main" java.io.IOError: java.io.IOException: Unable to create hard link from /raid0/cassandra/data/cassa_teads/stats_product-hc-233-Index.db to /raid0/cassandra/data/cassa_teads/snapshots/20120528/stats_product-hc-233-Index.db (errno 0) at org.apache.cassandra.db.ColumnFamilyStore.snapshotWithoutFlush(ColumnFamilyStore.java:1433) at org.apache.cassandra.db.ColumnFamilyStore.snapshot(ColumnFamilyStore.java:1462) at org.apache.cassandra.db.Table.snapshot(Table.java:210) at org.apache.cassandra.service.StorageService.takeSnapshot(StorageService.java:1710) at sun.reflect.NativeMethodAccessorImpl.invoke0(Native Method) at sun.reflect.NativeMethodAccessorImpl.invoke(NativeMethodAccessorImpl.java:39) at sun.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:25) at java.lang.reflect.Method.invoke(Method.java:597) at com.sun.jmx.mbeanserver.StandardMBeanIntrospector.invokeM2(StandardMBeanIntrospector.java:93) at com.sun.jmx.mbeanserver.StandardMBeanIntrospector.invokeM2(StandardMBeanIntrospector.java:27) at com.sun.jmx.mbeanserver.MBeanIntrospector.invokeM(MBeanIntrospector.java:208) at com.sun.jmx.mbeanserver.PerInterface.invoke(PerInterface.java:120) at com.sun.jmx.mbeanserver.MBeanSupport.invoke(MBeanSupport.java:262) at com.sun.jmx.interceptor.DefaultMBeanServerInterceptor.invoke(DefaultMBeanServerInterceptor.java:836) at com.sun.jmx.mbeanserver.JmxMBeanServer.invoke(JmxMBeanServer.java:761) at javax.management.remote.rmi.RMIConnectionImpl.doOperation(RMIConnectionImpl.java:1427) at javax.management.remote.rmi.RMIConnectionImpl.access$200(RMIConnectionImpl.java:72) at javax.management.remote.rmi.RMIConnectionImpl$PrivilegedOperation.run(RMIConnectionImpl.java:1265) at javax.management.remote.rmi.RMIConnectionImpl.doPrivilegedOperation(RMIConnectionImpl.java:1360) at javax.management.remote.rmi.RMIConnectionImpl.invoke(RMIConnectionImpl.java:788) at sun.reflect.GeneratedMethodAccessor50.invoke(Unknown Source) at sun.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:25) at java.lang.reflect.Method.invoke(Method.java:597) at sun.rmi.server.UnicastServerRef.dispatch(UnicastServerRef.java:303) at sun.rmi.transport.Transport$1.run(Transport.java:159) at java.security.AccessController.doPrivileged(Native Method) at sun.rmi.transport.Transport.serviceCall(Transport.java:155) at sun.rmi.transport.tcp.TCPTransport.handleMessages(TCPTransport.java:535) at sun.rmi.transport.tcp.TCPTransport$ConnectionHandler.run0(TCPTransport.java:790) at sun.rmi.transport.tcp.TCPTransport$ConnectionHandler.run(TCPTransport.java:649) at java.util.concurrent.ThreadPoolExecutor$Worker.runTask(ThreadPoolExecutor.java:886) at java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:908) at java.lang.Thread.run(Thread.java:662) Caused by: java.io.IOException: Unable to create hard link from /raid0/cassandra/data/cassa_teads/stats_product-hc-233-Index.db to /raid0/cassandra/data/cassa_teads/snapshots/20120528/stats_product-hc-233-Index.db (errno 0) at org.apache.cassandra.utils.CLibrary.createHardLink(CLibrary.java:158) at org.apache.cassandra.io.sstable.SSTableReader.createLinks(SSTableReader.java:857) at org.apache.cassandra.db.ColumnFamilyStore.snapshotWithoutFlush(ColumnFamilyStore.java:1412) ... 32 more Can we do a snapshot manually (like flushing and after copying all the file into the snapshot folder) ? Alain 2012/5/19 Jonathan Ellis : > When these bugs are fixed: > https://issues.apache.org/jira/secure/IssueNavigator.jspa?reset=true&jqlQuery=project+%3D+CASSANDRA+AND+fixVersion+%3D+%221.1.1%22+AND+resolution+%3D+Unresolved+ORDER+BY+due+ASC%2C+priority+DESC%2C+created+ASC&mode=hide > > On Wed, May 16, 2012 at 6:35 PM, Bryan Fernandez > wrote: >> Does anyone know when 1.1.1 will be released? >> >> Thanks. >> >> On Tue, May 15, 2012 at 5:40 PM, Brandon Williams wrote: >>> >>> Probably https://issues.apache.org/jira/browse/CASSANDRA-4230 >>> >>> On Tue, May 15, 2012 at 4:08 PM, Bryan Fernandez >>> wrote: >>> > Greetings, >>> > >>> > We recently upgraded from 1.0.8 to 1.1.0. Everything has been running >>> > fine >>> > with the exception of snapshots. When attempting to snapshot any of the >>> > nodes in our six node cluster we are seeing the following error. >>> > >>> > [root@cassandra-n6 blotter]# /opt/apache-cas
Doubt in Row key range scan
Dear all I have stored my data into Cassandra database in the format "tickerID_date". Now when I specify the row key range like 1_2012/05/24(start) to 1_2012/05/27(end) it says that the end key md5 value is lesser than start key md5 value. So I changed my start key to 1_2012/05/27 and end key to 1_2012/05/24, then I got all the keys even which are not in my range like 67_2012/05/23 and 54_2012/05/28. I am using Thrift API. Please help me as I want only the columns of 1_2012/05/24, 1_2012/05/25 , 1_2012/05/26 and 1_2012/05/27. Prakrati Agrawal | Developer - Big Data(I&D)| 9731648376 | www.mu-sigma.com This email message may contain proprietary, private and confidential information. The information transmitted is intended only for the person(s) or entities to which it is addressed. Any review, retransmission, dissemination or other use of, or taking of any action in reliance upon, this information by persons or entities other than the intended recipient is prohibited and may be illegal. If you received this in error, please contact the sender and delete the message from your system. Mu Sigma takes all reasonable steps to ensure that its electronic communications are free from viruses. However, given Internet accessibility, the Company cannot accept liability for any virus introduced by this e-mail or any attachment and you are advised to use up-to-date virus checking software.
Re: Doubt in Row key range scan
Hi, It's normal. Keys to replicas are determined with a hash (md5) when using the random partitionner (which you are using I guess). You probably want to switch to the order preserving partionner or tweak your data model in order to rely on 2nd index for such filtering. - Pierre -Original Message- From: Prakrati Agrawal Date: Mon, 28 May 2012 04:39:46 To: user@cassandra.apache.org Reply-To: user@cassandra.apache.org Subject: Doubt in Row key range scan Dear all I have stored my data into Cassandra database in the format "tickerID_date". Now when I specify the row key range like 1_2012/05/24(start) to 1_2012/05/27(end) it says that the end key md5 value is lesser than start key md5 value. So I changed my start key to 1_2012/05/27 and end key to 1_2012/05/24, then I got all the keys even which are not in my range like 67_2012/05/23 and 54_2012/05/28. I am using Thrift API. Please help me as I want only the columns of 1_2012/05/24, 1_2012/05/25 , 1_2012/05/26 and 1_2012/05/27. Prakrati Agrawal | Developer - Big Data(I&D)| 9731648376 | www.mu-sigma.com This email message may contain proprietary, private and confidential information. The information transmitted is intended only for the person(s) or entities to which it is addressed. Any review, retransmission, dissemination or other use of, or taking of any action in reliance upon, this information by persons or entities other than the intended recipient is prohibited and may be illegal. If you received this in error, please contact the sender and delete the message from your system. Mu Sigma takes all reasonable steps to ensure that its electronic communications are free from viruses. However, given Internet accessibility, the Company cannot accept liability for any virus introduced by this e-mail or any attachment and you are advised to use up-to-date virus checking software.
Re: Doubt in Row key range scan
You are using the Random Partitioner. Using the RP is a good thing because you avoid hot spots, but it has its defaults too. You can't scan a slice of row, they won't be ordered because all your keys are stored using their md5 values. You should review your data model to use columns to order your data. Alain 2012/5/28 Prakrati Agrawal : > Dear all > > > > I have stored my data into Cassandra database in the format “tickerID_date”. > Now when I specify the row key range like 1_2012/05/24(start) to > 1_2012/05/27(end) it says that the end key md5 value is lesser than start > key md5 value. So I changed my start key to 1_2012/05/27 and end key to > 1_2012/05/24, then I got all the keys even which are not in my range like > 67_2012/05/23 and 54_2012/05/28. I am using Thrift API. > > Please help me as I want only the columns of 1_2012/05/24, 1_2012/05/25 , > 1_2012/05/26 and 1_2012/05/27. > > > > Prakrati Agrawal | Developer - Big Data(I&D)| 9731648376 | www.mu-sigma.com > > > > > > This email message may contain proprietary, private and confidential > information. The information transmitted is intended only for the person(s) > or entities to which it is addressed. Any review, retransmission, > dissemination or other use of, or taking of any action in reliance upon, > this information by persons or entities other than the intended recipient is > prohibited and may be illegal. If you received this in error, please contact > the sender and delete the message from your system. > > Mu Sigma takes all reasonable steps to ensure that its electronic > communications are free from viruses. However, given Internet accessibility, > the Company cannot accept liability for any virus introduced by this e-mail > or any attachment and you are advised to use up-to-date virus checking > software.
RE: Doubt in Row key range scan
Please could you tell me how to tweak my data model to rely on 2nd index ? Thank you Prakrati Agrawal | Developer - Big Data(I&D)| 9731648376 | www.mu-sigma.com From: Pierre Chalamet [mailto:pie...@chalamet.net] Sent: Monday, May 28, 2012 3:31 PM To: user@cassandra.apache.org Subject: Re: Doubt in Row key range scan Hi, It's normal. Keys to replicas are determined with a hash (md5) when using the random partitionner (which you are using I guess). You probably want to switch to the order preserving partionner or tweak your data model in order to rely on 2nd index for such filtering. - Pierre From: Prakrati Agrawal Date: Mon, 28 May 2012 04:39:46 -0500 To: user@cassandra.apache.org ReplyTo: user@cassandra.apache.org Subject: Doubt in Row key range scan Dear all I have stored my data into Cassandra database in the format "tickerID_date". Now when I specify the row key range like 1_2012/05/24(start) to 1_2012/05/27(end) it says that the end key md5 value is lesser than start key md5 value. So I changed my start key to 1_2012/05/27 and end key to 1_2012/05/24, then I got all the keys even which are not in my range like 67_2012/05/23 and 54_2012/05/28. I am using Thrift API. Please help me as I want only the columns of 1_2012/05/24, 1_2012/05/25 , 1_2012/05/26 and 1_2012/05/27. Prakrati Agrawal | Developer - Big Data(I&D)| 9731648376 | www.mu-sigma.com This email message may contain proprietary, private and confidential information. The information transmitted is intended only for the person(s) or entities to which it is addressed. Any review, retransmission, dissemination or other use of, or taking of any action in reliance upon, this information by persons or entities other than the intended recipient is prohibited and may be illegal. If you received this in error, please contact the sender and delete the message from your system. Mu Sigma takes all reasonable steps to ensure that its electronic communications are free from viruses. However, given Internet accessibility, the Company cannot accept liability for any virus introduced by this e-mail or any attachment and you are advised to use up-to-date virus checking software. This email message may contain proprietary, private and confidential information. The information transmitted is intended only for the person(s) or entities to which it is addressed. Any review, retransmission, dissemination or other use of, or taking of any action in reliance upon, this information by persons or entities other than the intended recipient is prohibited and may be illegal. If you received this in error, please contact the sender and delete the message from your system. Mu Sigma takes all reasonable steps to ensure that its electronic communications are free from viruses. However, given Internet accessibility, the Company cannot accept liability for any virus introduced by this e-mail or any attachment and you are advised to use up-to-date virus checking software.
project setup in asp.net Fluent Cassandra
Hi All, Right now i am trying to use Fluent Cassandra API for development in asp.net . For the same purpose i have downloaded the same source from github. But every time i add a new reference *fluentcassandra.dll*, i am unable to use it in my code. I am not able figure it out why this is happening? Any one has any clue. Thanks In Advance, Abhijit
Re: Doubt in Row key range scan
Check this out: http://www.anuff.com/2011/02/indexing-in-cassandra.html#more Or just google for wide row indexes. On May 28, 2012, at 11:22 AM, Prakrati Agrawal wrote: > Please could you tell me how to tweak my data model to rely on 2nd index ? > Thank you > > > Prakrati Agrawal | Developer - Big Data(I&D)| 9731648376 | www.mu-sigma.com > > From: Pierre Chalamet [mailto:pie...@chalamet.net] > Sent: Monday, May 28, 2012 3:31 PM > To: user@cassandra.apache.org > Subject: Re: Doubt in Row key range scan > > Hi, > > It's normal. > > Keys to replicas are determined with a hash (md5) when using the random > partitionner (which you are using I guess). > > You probably want to switch to the order preserving partionner or tweak your > data model in order to rely on 2nd index for such filtering. > - Pierre > From: Prakrati Agrawal > Date: Mon, 28 May 2012 04:39:46 -0500 > To: user@cassandra.apache.org > ReplyTo: user@cassandra.apache.org > Subject: Doubt in Row key range scan > > Dear all > > I have stored my data into Cassandra database in the format “tickerID_date”. > Now when I specify the row key range like 1_2012/05/24(start) to > 1_2012/05/27(end) it says that the end key md5 value is lesser than start key > md5 value. So I changed my start key to 1_2012/05/27 and end key to > 1_2012/05/24, then I got all the keys even which are not in my range like > 67_2012/05/23 and 54_2012/05/28. I am using Thrift API. > Please help me as I want only the columns of 1_2012/05/24, 1_2012/05/25 , > 1_2012/05/26 and 1_2012/05/27. > > Prakrati Agrawal | Developer - Big Data(I&D)| 9731648376 | www.mu-sigma.com > > > This email message may contain proprietary, private and confidential > information. The information transmitted is intended only for the person(s) > or entities to which it is addressed. Any review, retransmission, > dissemination or other use of, or taking of any action in reliance upon, this > information by persons or entities other than the intended recipient is > prohibited and may be illegal. If you received this in error, please contact > the sender and delete the message from your system. > > Mu Sigma takes all reasonable steps to ensure that its electronic > communications are free from viruses. However, given Internet accessibility, > the Company cannot accept liability for any virus introduced by this e-mail > or any attachment and you are advised to use up-to-date virus checking > software. > > This email message may contain proprietary, private and confidential > information. The information transmitted is intended only for the person(s) > or entities to which it is addressed. Any review, retransmission, > dissemination or other use of, or taking of any action in reliance upon, this > information by persons or entities other than the intended recipient is > prohibited and may be illegal. If you received this in error, please contact > the sender and delete the message from your system. > > Mu Sigma takes all reasonable steps to ensure that its electronic > communications are free from viruses. However, given Internet accessibility, > the Company cannot accept liability for any virus introduced by this e-mail > or any attachment and you are advised to use up-to-date virus checking > software. Cumprimentos, Luís Ferreira
commitlog_sync_batch_window_in_ms change in 0.7
I'm experimenting with Cassandra 0.7 for some time now. I want to increase the value of commitlog_sync_batch_window_in_ms, without losing previous data. I can shut down my cluster, but I don't want to have the nodes ignore it due to some schema misoconfiguration etc when I get it up again. I apologize if this was asked before, but I did not see a clear guide for achieving something like this. Can someone please help? Thank you
Re: commitlog_sync_batch_window_in_ms change in 0.7
Hi, Using nodetool for each node one by one: 1. disablethrift 2. drain 3. Shutdown your daemon 4. Modify the config 5. Restart the node You won't lose the data on your nodes - clients might see a node down, it is usually not a problem if your c* client is smart enough. You also won't lose updates while performing your operations if the cl is quorum (and have enough replicas). --Original Message-- From: osishkin osishkin To: user@cassandra.apache.org ReplyTo: user@cassandra.apache.org ReplyTo: osish...@gmail.com Subject: commitlog_sync_batch_window_in_ms change in 0.7 Sent: May 28, 2012 15:53 I'm experimenting with Cassandra 0.7 for some time now. I want to increase the value of commitlog_sync_batch_window_in_ms, without losing previous data. I can shut down my cluster, but I don't want to have the nodes ignore it due to some schema misoconfiguration etc when I get it up again. I apologize if this was asked before, but I did not see a clear guide for achieving something like this. Can someone please help? Thank you - Pierre
Doubts regarding compaction
http://www.datastax.com/docs/1.0/operations/tuning On this page at last, there's a note about Major Compaction which says, Also, once you run a major compaction, automatic minor compactions are no > longer triggered frequently... Could anybody give an explanation for that, because as far as I think, once a major compaction takes place, after that let's say there would be no compactions till N(default value 4) SSTables of same size (size of memtable to be precise) are formed, then automatically minor compactions would start. Is the author referring to this time period as no minor compactions being triggered automatically ? Also, I want to make sure, if Major compactions could only be done manually ? -- *Rajat Mathur*
Concurrency Control
Hi, what kind of Concurrency Control Method is used in Cassandra? I found out so far that it's not done with the MVCC Method and that no vector clocks are being used. Thanks Helen