Hello,

I am trying to run a Hadoop job that pulls data out of Cassandra via 
ColumnFamilyInputFormat. I am getting a "frame size" exception. To remedy that, 
I have set both the thrift_framed_transport_size_in_mb and 
thrift_max_message_length_in_mb to an "infinite" amount at 100000mb on all 
nodes. Moreover, I have restarted the cluster and the cassandra.yaml files have 
been reloaded.
However, I am still getting:

12/11/09 21:39:52 INFO mapred.JobClient:  map 62% reduce 0%
12/11/09 21:40:09 INFO mapred.JobClient: Task Id : 
attempt_201211082011_0015_m_000479_2, Status : FAILED
java.lang.RuntimeException: org.apache.thrift.transport.TTransportException: 
Frame size (30046945) larger than max length (16384000)!
        at 
org.apache.cassandra.hadoop.ColumnFamilyRecordReader$StaticRowIterator.maybeInit(ColumnFamilyRecordReader.java:400)
        at 
org.apache.cassandra.hadoop.ColumnFamilyRecordReader$StaticRowIterator.computeNext(ColumnFamilyRecordReader.java:406)
        at 
org.apache.cassandra.hadoop.ColumnFamilyRecordReader$StaticRowIterator.computeNext(ColumnFamilyRecordReader.java:324)
        at 
com.google.common.collect.AbstractIterator.tryToComputeNext(AbstractIterator.java:143)
        at 
com.google.common.collect.AbstractIterator.hasNext(AbstractIterator.java:138)
        at 
org.apache.cassandra.hadoop.ColumnFamilyRecordReader.nextKeyValue(ColumnFamilyRecordReader.java:189)

Question: Why is 16384000 bytes (I assume) !=  100000mb?

Next, I made this parameter true as a last hail mary attempt:
        cassandra.input.widerows=true
...still with no luck.

Does someone know what I might be missing?

Thank you very much for your time,
Marko.

http://markorodriguez.com

Reply via email to