Re: newer Cassandra + Hadoop = TimedOutException()

2012-03-08 Thread Patrik Modesto
I did change the rpc_endpoint to endpoints and now the splits are computed correctly. So it's a bug in cassandra to hadoop interface. I suspect that it has something to do with wide rows with tens of thousands of columns we have because the unpatched getSubSplits() works with small test data we hav

Re: newer Cassandra + Hadoop = TimedOutException()

2012-03-07 Thread Florent Lefillâtre
If you want try a test, in the CFIF.getSubSplits(String, String, TokenRange, Configuration) method, replace the loop on 'range.rpc_endpoints' by the same loop on 'range.endpoints'. This method split token range of each node with describe_splits method, but I think there is something wrong when you

Re: newer Cassandra + Hadoop = TimedOutException()

2012-03-07 Thread Patrik Modesto
You're right, I wasn't looking in the right logs. Unfortunately I'd need to restart hadoop takstracker with loglevel DEBUG and that is not possilbe at the moment. Pitty it happens only in the production with terrabytes of data, not in the test... Regards, P. On Tue, Mar 6, 2012 at 14:31, Florent

Re: newer Cassandra + Hadoop = TimedOutException()

2012-03-06 Thread Florent Lefillâtre
CFRR.getProgress() is called by child mapper tasks on each TastTracker node, so the log must appear on ${hadoop_log_dir}/attempt_201202081707_0001_m_00_0/syslog (or somethings like this) on TaskTrackers, not on client job logs. Are you sure to see the good log file, I say that because in your f

Re: newer Cassandra + Hadoop = TimedOutException()

2012-03-06 Thread Patrik Modesto
I've added a debug message in the CFRR.getProgress() and I can't find it in the debug output. Seems like the getProgress() has not been called at all; Regards, P. On Tue, Mar 6, 2012 at 09:49, Jeremy Hanna wrote: > you may be running into this - > https://issues.apache.org/jira/browse/CASSANDRA

Re: newer Cassandra + Hadoop = TimedOutException()

2012-03-06 Thread Patrik Modesto
I've tryied cassandra-all 0.8.10 with fixed the rpc_endpoints == "0.0.0.0" bug, but the result is the same, there are still tasks over 1000%. The only change is that there are real host names instead of 0.0.0.0 in the debug output. Reconfiguring whole cluster is not possible, I can't test the "rp

Re: newer Cassandra + Hadoop = TimedOutException()

2012-03-06 Thread Florent Lefillâtre
I remember a bug on the ColumnFamilyInputFormat class 0.8.10. It was a test rpc_endpoints == "0.0.0.0" in place of rpc_endpoint.equals("0.0.0.0"), may be it can help you Le 6 mars 2012 12:18, Florent Lefillâtre a écrit : > Excuse me, I had not understood. > So, for me, the problem comes from the

Re: newer Cassandra + Hadoop = TimedOutException()

2012-03-06 Thread Florent Lefillâtre
Excuse me, I had not understood. So, for me, the problem comes from the change of ColumnFamilyInputFormat class between 0.8.7 and 0.8.10 where the splits are created (0.8.7 uses endpoints and 0.8.10 uses rpc_endpoints). With your config, splits fails, so Hadoop doesn't run a Map task on approximtiv

Re: newer Cassandra + Hadoop = TimedOutException()

2012-03-06 Thread Patrik Modesto
Hi Florent, I don't change the server version, it is the Cassandra 0.8.10. I change just the version of cassandra-all in pom.xml of the mapreduce job. I have the 'rpc_address: 0.0.0.0' in cassandra.yaml, because I want cassandra to bind RPC to all interfaces. Regards, P. On Tue, Mar 6, 2012 at

Re: newer Cassandra + Hadoop = TimedOutException()

2012-03-06 Thread Jeremy Hanna
you may be running into this - https://issues.apache.org/jira/browse/CASSANDRA-3942 - I'm not sure if it really affects the execution of the job itself though. On Mar 6, 2012, at 2:32 AM, Patrik Modesto wrote: > Hi, > > I was recently trying Hadoop job + cassandra-all 0.8.10 again and the > Ti

Re: newer Cassandra + Hadoop = TimedOutException()

2012-03-06 Thread Florent Lefillâtre
Hi, I had the same problem on hadoop 0.20.2 and cassandra 1.0.5. In my case the split of token range failed. I have comment line 'rpc_address: 0.0.0.0' in cassandra.yaml. May be see if you have not configuration changes between 0.8.7 and 0.8.10 Le 6 mars 2012 09:32, Patrik Modesto a écrit : > H

Re: newer Cassandra + Hadoop = TimedOutException()

2012-03-06 Thread Patrik Modesto
Hi, I was recently trying Hadoop job + cassandra-all 0.8.10 again and the Timeouts I get are not because of the Cassandra can't handle the requests. I've noticed there are several tasks that show proggess of several thousands percents. Seems like they are looping their range of keys. I've run the

Re: newer Cassandra + Hadoop = TimedOutException()

2012-02-28 Thread Patrik Modesto
I'll alter these settings and will let you know. Regards, P. On Tue, Feb 28, 2012 at 09:23, aaron morton wrote: > Have you tried lowering the  batch size and increasing the time out? Even > just to get it to work. > > If you get a TimedOutException it means CL number of servers did not respond >

Re: newer Cassandra + Hadoop = TimedOutException()

2012-02-28 Thread aaron morton
Have you tried lowering the batch size and increasing the time out? Even just to get it to work. If you get a TimedOutException it means CL number of servers did not respond in time. Cheers - Aaron Morton Freelance Developer @aaronmorton http://www.thelastpickle.com On 28/0

Re: newer Cassandra + Hadoop = TimedOutException()

2012-02-27 Thread Patrik Modesto
Hi aaron, this is our current settings: cassandra.range.batch.size 1024 cassandra.input.split.size 16384 rpc_timeout_in_ms: 3 Regards, P. On Mon, Feb 27, 2012 at 21:54, aaron morton wrote: > What settings do you have for

Re: newer Cassandra + Hadoop = TimedOutException()

2012-02-27 Thread aaron morton
What settings do you have for cassandra.range.batch.size and rpc_timeout_in_ms ? Have you tried reducing the first and/or increasing the second ? Cheers - Aaron Morton Freelance Developer @aaronmorton http://www.thelastpickle.com On 27/02/2012, at 8:02 PM, Patrik Modesto wrote

Re: newer Cassandra + Hadoop = TimedOutException()

2012-02-26 Thread Patrik Modesto
On Sun, Feb 26, 2012 at 04:25, Edward Capriolo wrote: > Did you see the notes here? I'm not sure what do you mean by the notes? I'm using the mapred.* settings suggested there: mapred.max.tracker.failures 20 mapred.map.max.attempts 20

Re: newer Cassandra + Hadoop = TimedOutException()

2012-02-25 Thread Edward Capriolo
Did you see the notes here? http://wiki.apache.org/cassandra/HadoopSupport#Troubleshooting On Sat, Feb 25, 2012 at 2:04 AM, Patrik Modesto wrote: > No, we have our own servers, no cloud service. > > P. > > On Feb 24, 2012 5:57 PM, "Jeremy Hanna" wrote: >> >> By chance are you in EC2? >> >> On F

Re: newer Cassandra + Hadoop = TimedOutException()

2012-02-24 Thread Patrik Modesto
No, we have our own servers, no cloud service. P. On Feb 24, 2012 5:57 PM, "Jeremy Hanna" wrote: > By chance are you in EC2? > > On Feb 24, 2012, at 8:33 AM, Patrik Modesto wrote: > > > Hi Jeremy, > > > > I've seen the page and tried the values but to no help. > > > > Here goes tcpdump of one fa

Re: newer Cassandra + Hadoop = TimedOutException()

2012-02-24 Thread Jeremy Hanna
By chance are you in EC2? On Feb 24, 2012, at 8:33 AM, Patrik Modesto wrote: > Hi Jeremy, > > I've seen the page and tried the values but to no help. > > Here goes tcpdump of one failed TCP connection: > > 15:06:20.231421 IP 10.0.18.87.9160 > 10.0.18.87.39396: Flags [P.], seq > 137891735:13790

Re: newer Cassandra + Hadoop = TimedOutException()

2012-02-24 Thread Patrik Modesto
Hi Jeremy, I've seen the page and tried the values but to no help. Here goes tcpdump of one failed TCP connection: 15:06:20.231421 IP 10.0.18.87.9160 > 10.0.18.87.39396: Flags [P.], seq 137891735:137904068, ack 708, win 282, options [nop,nop,TS val 4119981636 ecr 4119981636], length 12333 15:06:

Re: newer Cassandra + Hadoop = TimedOutException()

2012-02-24 Thread Jeremy Hanna
Check out the troubleshooting section of the hadoop support - we ran into the same thing and tried to update that with some info on how to get around it: http://wiki.apache.org/cassandra/HadoopSupport#Troubleshooting On Feb 24, 2012, at 7:20 AM, Patrik Modesto wrote: > Hi, > > I can see some st

Re: newer Cassandra + Hadoop = TimedOutException()

2012-02-24 Thread Patrik Modesto
There are two more similar exceptions: DEBUG 14:23:12,817 Thrift transport error occurred during processing of message. org.apache.thrift.transport.TTransportException at org.apache.thrift.transport.TIOStreamTransport.read(TIOStreamTransport.java:132) at org.apache.thrift.transpor

Re: newer Cassandra + Hadoop = TimedOutException()

2012-02-24 Thread Patrik Modesto
Some more info, running the cluster with debug enabled shows this error: DEBUG 14:23:05,749 Thrift transport error occurred during processing of message. org.apache.thrift.transport.TTransportException: java.net.SocketException: Broken pipe at org.apache.thrift.transport.TIOStreamTranspor