ReliableKafkaReceiver stopped receiving data after WriteAheadLogBasedBlockHandler throws TimeoutException

2015-01-12 Thread Max Xu
Hi all, I am running a Spark streaming application with ReliableKafkaReceiver (Spark 1.2.0). Constantly I was getting the following exception: 15/01/12 19:07:06 ERROR receiver.BlockGenerator: Error in block pushing thread java.util.concurrent.TimeoutException: Futures timed out after [30 seconds

streaming application throws IOException due to Log directory already exists during checkpoint recovery

2015-01-07 Thread Max Xu
Hi All, I run a Spark streaming application (Spark 1.2.0) on YARN (Hadoop 2.5.2) with Spark event log enabled. I set the checkpoint dir in the streaming context and run the app. It started in YARN with application id 'app_id_1' and created the Spark event log dir /spark/applicationHistory/app_i

RE: Saving data to Hbase hung in Spark streaming application with Spark 1.2.0

2015-01-06 Thread Max Xu
, January 06, 2015 1:12 PM To: Max Xu Cc: user@spark.apache.org Subject: Re: Saving data to Hbase hung in Spark streaming application with Spark 1.2.0 I doubt anyone would deploy hbase 0.98.x on hadoop-1 Looks like hadoop2 profile should be made the default. Cheers On Tue, Jan 6, 2015 at 9:49 AM

RE: Saving data to Hbase hung in Spark streaming application with Spark 1.2.0

2015-01-06 Thread Max Xu
0.98.7-hadoop1 From: Ted Yu [mailto:yuzhih...@gmail.com] Sent: Tuesday, January 06, 2015 12:39 PM To: Max Xu Cc: user@spark.apache.org Subject: Re: Saving data to Hbase hung in Spark streaming application with Spark 1.2.0 Default profile is hbase-hadoop1 so you need to specify

RE: Saving data to Hbase hung in Spark streaming application with Spark 1.2.0

2015-01-06 Thread Max Xu
use spark-examples-1.1.1-hadoop2.5.2.jar from Spark 1.1.1 build (build with hbase 0.98.8-hadoop2 ), the problem went away. I’ll try to run the app again after rebuild Spark 1.2.0 with 0.98.8-hadoop2. From: Ted Yu [mailto:yuzhih...@gmail.com] Sent: Tuesday, January 06, 2015 11:56 AM To: Max Xu Cc

Saving data to Hbase hung in Spark streaming application with Spark 1.2.0

2015-01-06 Thread Max Xu
Hi all, I have a Spark streaming application that ingests data from a Kafka topic and persists received data to Hbase. It works fine with Spark 1.1.1 in YARN cluster mode. Basically, I use the following code to persist each partition of each RDD to Hbase: @Override void call(It