Thanks Sebastian, I will try it out. But I am also curious why is the COPY command failing with Out of Memory Error.
regards Neha On Tue, Apr 21, 2015 at 4:35 AM, Sebastian Estevez < sebastian.este...@datastax.com> wrote: > Blobs are ByteBuffer s it calls getBytes().toString: > > > https://github.com/brianmhess/cassandra-loader/blob/master/src/main/java/com/datastax/loader/parser/ByteBufferParser.java#L35 > > All the best, > > > [image: datastax_logo.png] <http://www.datastax.com/> > > Sebastián Estévez > > Solutions Architect | 954 905 8615 | sebastian.este...@datastax.com > > [image: linkedin.png] <https://www.linkedin.com/company/datastax> [image: > facebook.png] <https://www.facebook.com/datastax> [image: twitter.png] > <https://twitter.com/datastax> [image: g+.png] > <https://plus.google.com/+Datastax/about> > <http://feeds.feedburner.com/datastax> > > <http://cassandrasummit-datastax.com/> > > DataStax is the fastest, most scalable distributed database technology, > delivering Apache Cassandra to the world’s most innovative enterprises. > Datastax is built to be agile, always-on, and predictably scalable to any > size. With more than 500 customers in 45 countries, DataStax is the > database technology and transactional backbone of choice for the worlds > most innovative companies such as Netflix, Adobe, Intuit, and eBay. > > On Mon, Apr 20, 2015 at 5:47 PM, Serega Sheypak <serega.shey...@gmail.com> > wrote: > >> hi, what happens if unloader meets blob field? >> >> 2015-04-20 23:43 GMT+02:00 Sebastian Estevez < >> sebastian.este...@datastax.com>: >> >>> Try Brian's cassandra-unloader >>> <https://github.com/brianmhess/cassandra-loader#cassandra-unloader> >>> >>> All the best, >>> >>> >>> [image: datastax_logo.png] <http://www.datastax.com/> >>> >>> Sebastián Estévez >>> >>> Solutions Architect | 954 905 8615 | sebastian.este...@datastax.com >>> >>> [image: linkedin.png] <https://www.linkedin.com/company/datastax> [image: >>> facebook.png] <https://www.facebook.com/datastax> [image: twitter.png] >>> <https://twitter.com/datastax> [image: g+.png] >>> <https://plus.google.com/+Datastax/about> >>> <http://feeds.feedburner.com/datastax> >>> >>> <http://cassandrasummit-datastax.com/> >>> >>> DataStax is the fastest, most scalable distributed database technology, >>> delivering Apache Cassandra to the world’s most innovative enterprises. >>> Datastax is built to be agile, always-on, and predictably scalable to any >>> size. With more than 500 customers in 45 countries, DataStax is the >>> database technology and transactional backbone of choice for the worlds >>> most innovative companies such as Netflix, Adobe, Intuit, and eBay. >>> >>> On Mon, Apr 20, 2015 at 12:31 PM, Neha Trivedi <nehajtriv...@gmail.com> >>> wrote: >>> >>>> Does the nproc,nofile,memlock settings in >>>> /etc/security/limits.d/cassandra.conf are set to optimum value ? >>>> it's all default. >>>> >>>> What is the consistency level ? >>>> CL = Qurom >>>> >>>> Is there any other way to export a table to CSV? >>>> >>>> regards >>>> Neha >>>> >>>> On Mon, Apr 20, 2015 at 12:21 PM, Kiran mk <coolkiran2...@gmail.com> >>>> wrote: >>>> >>>>> Hi, >>>>> >>>>> Thanks for the info, >>>>> >>>>> Does the nproc,nofile,memlock settings in >>>>> /etc/security/limits.d/cassandra.conf are set to optimum value ? >>>>> >>>>> What is the consistency level ? >>>>> >>>>> Best Regardds, >>>>> Kiran.M.K. >>>>> >>>>> >>>>> On Mon, Apr 20, 2015 at 11:55 AM, Neha Trivedi <nehajtriv...@gmail.com >>>>> > wrote: >>>>> >>>>>> hi, >>>>>> >>>>>> What is the count of records in the column-family ? >>>>>> We have about 38,000 Rows in the column-family for which we are >>>>>> trying to export >>>>>> What is the Cassandra Version ? >>>>>> We are using Cassandra 2.0.11 >>>>>> >>>>>> MAX_HEAP_SIZE and HEAP_NEWSIZE is the default . >>>>>> The Server is 8 GB. >>>>>> >>>>>> regards >>>>>> Neha >>>>>> >>>>>> On Mon, Apr 20, 2015 at 11:39 AM, Kiran mk <coolkiran2...@gmail.com> >>>>>> wrote: >>>>>> >>>>>>> Hi, >>>>>>> >>>>>>> check the MAX_HEAP_SIZE configuration in cassandra-env.sh >>>>>>> environment file >>>>>>> >>>>>>> Also HEAP_NEWSIZE ? >>>>>>> >>>>>>> What is the Consistency Level you are using ? >>>>>>> >>>>>>> Best REgards, >>>>>>> Kiran.M.K. >>>>>>> >>>>>>> On Mon, Apr 20, 2015 at 11:13 AM, Kiran mk <coolkiran2...@gmail.com> >>>>>>> wrote: >>>>>>> >>>>>>>> Seems like the is related to JAVA HEAP Memory. >>>>>>>> >>>>>>>> What is the count of records in the column-family ? >>>>>>>> >>>>>>>> What is the Cassandra Version ? >>>>>>>> >>>>>>>> Best Regards, >>>>>>>> Kiran.M.K. >>>>>>>> >>>>>>>> On Mon, Apr 20, 2015 at 11:08 AM, Neha Trivedi < >>>>>>>> nehajtriv...@gmail.com> wrote: >>>>>>>> >>>>>>>>> Hello all, >>>>>>>>> >>>>>>>>> We are getting the OutOfMemoryError on one of the Node and the >>>>>>>>> Node is down, when we run the export command to get all the data from >>>>>>>>> a >>>>>>>>> table. >>>>>>>>> >>>>>>>>> >>>>>>>>> Regards >>>>>>>>> Neha >>>>>>>>> >>>>>>>>> >>>>>>>>> >>>>>>>>> >>>>>>>>> ERROR [ReadStage:532074] 2015-04-09 01:04:00,603 >>>>>>>>> CassandraDaemon.java (line 199) Exception in thread >>>>>>>>> Thread[ReadStage:532074,5,main] >>>>>>>>> java.lang.OutOfMemoryError: Java heap space >>>>>>>>> at >>>>>>>>> org.apache.cassandra.io.util.RandomAccessReader.readBytes(RandomAccessReader.java:347) >>>>>>>>> at >>>>>>>>> org.apache.cassandra.utils.ByteBufferUtil.read(ByteBufferUtil.java:392) >>>>>>>>> at >>>>>>>>> org.apache.cassandra.utils.ByteBufferUtil.readWithLength(ByteBufferUtil.java:355) >>>>>>>>> at >>>>>>>>> org.apache.cassandra.db.ColumnSerializer.deserializeColumnBody(ColumnSerializer.java:124) >>>>>>>>> at >>>>>>>>> org.apache.cassandra.db.OnDiskAtom$Serializer.deserializeFromSSTable(OnDiskAtom.java:85) >>>>>>>>> at >>>>>>>>> org.apache.cassandra.db.Column$1.computeNext(Column.java:75) >>>>>>>>> at >>>>>>>>> org.apache.cassandra.db.Column$1.computeNext(Column.java:64) >>>>>>>>> at >>>>>>>>> com.google.common.collect.AbstractIterator.tryToComputeNext(AbstractIterator.java:143) >>>>>>>>> at >>>>>>>>> com.google.common.collect.AbstractIterator.hasNext(AbstractIterator.java:138) >>>>>>>>> at >>>>>>>>> org.apache.cassandra.db.columniterator.SimpleSliceReader.computeNext(SimpleSliceReader.java:88) >>>>>>>>> at >>>>>>>>> org.apache.cassandra.db.columniterator.SimpleSliceReader.computeNext(SimpleSliceReader.java:37) >>>>>>>>> at >>>>>>>>> com.google.common.collect.AbstractIterator.tryToComputeNext(AbstractIterator.java:143) >>>>>>>>> at >>>>>>>>> com.google.common.collect.AbstractIterator.hasNext(AbstractIterator.java:138) >>>>>>>>> at >>>>>>>>> org.apache.cassandra.db.columniterator.SSTableSliceIterator.hasNext(SSTableSliceIterator.java:82) >>>>>>>>> at >>>>>>>>> org.apache.cassandra.db.columniterator.LazyColumnIterator.computeNext(LazyColumnIterator.java:82) >>>>>>>>> at >>>>>>>>> org.apache.cassandra.db.columniterator.LazyColumnIterator.computeNext(LazyColumnIterator.java:59) >>>>>>>>> at >>>>>>>>> com.google.common.collect.AbstractIterator.tryToComputeNext(AbstractIterator.java:143) >>>>>>>>> at >>>>>>>>> com.google.common.collect.AbstractIterator.hasNext(AbstractIterator.java:138) >>>>>>>>> at >>>>>>>>> org.apache.cassandra.db.filter.QueryFilter$2.getNext(QueryFilter.java:157) >>>>>>>>> at >>>>>>>>> org.apache.cassandra.db.filter.QueryFilter$2.hasNext(QueryFilter.java:140) >>>>>>>>> at >>>>>>>>> org.apache.cassandra.utils.MergeIterator$OneToOne.computeNext(MergeIterator.java:200) >>>>>>>>> at >>>>>>>>> com.google.common.collect.AbstractIterator.tryToComputeNext(AbstractIterator.java:143) >>>>>>>>> at >>>>>>>>> com.google.common.collect.AbstractIterator.hasNext(AbstractIterator.java:138) >>>>>>>>> at >>>>>>>>> org.apache.cassandra.db.filter.SliceQueryFilter.collectReducedColumns(SliceQueryFilter.java:185) >>>>>>>>> at >>>>>>>>> org.apache.cassandra.db.filter.QueryFilter.collateColumns(QueryFilter.java:122) >>>>>>>>> at >>>>>>>>> org.apache.cassandra.db.filter.QueryFilter.collateOnDiskAtom(QueryFilter.java:80) >>>>>>>>> at >>>>>>>>> org.apache.cassandra.db.RowIteratorFactory$2.getReduced(RowIteratorFactory.java:101) >>>>>>>>> at >>>>>>>>> org.apache.cassandra.db.RowIteratorFactory$2.getReduced(RowIteratorFactory.java:75) >>>>>>>>> at >>>>>>>>> org.apache.cassandra.utils.MergeIterator$ManyToOne.consume(MergeIterator.java:115) >>>>>>>>> at >>>>>>>>> org.apache.cassandra.utils.MergeIterator$ManyToOne.computeNext(MergeIterator.java:98) >>>>>>>>> at >>>>>>>>> com.google.common.collect.AbstractIterator.tryToComputeNext(AbstractIterator.java:143) >>>>>>>>> at >>>>>>>>> com.google.common.collect.AbstractIterator.hasNext(AbstractIterator.java:138) >>>>>>>>> >>>>>>>>> >>>>>>>> >>>>>>>> >>>>>>>> >>>>>>>> -- >>>>>>>> Best Regards, >>>>>>>> Kiran.M.K. >>>>>>>> >>>>>>> >>>>>>> >>>>>>> >>>>>>> -- >>>>>>> Best Regards, >>>>>>> Kiran.M.K. >>>>>>> >>>>>> >>>>>> >>>>> >>>>> >>>>> -- >>>>> Best Regards, >>>>> Kiran.M.K. >>>>> >>>> >>>> >>> >> >