rdd.take() failed when input partition is larger than hdfs blocksize

2014-03-10 Thread Chen Jin
Hello Spark Developers, While trying to use rdd.take(numItems) My job just hangs there forever, the following are output messages: 14/03/07 00:52:21 INFO SparkContext: Starting job: take at xx.java:55 14/03/07 00:52:21 INFO DAGScheduler: Got job 1 (take at xx.java:55) with 1 output

rdd.take() failed when input partition is larger than hdfs blocksize

2014-03-06 Thread Chen Jin
Hello Spark Developers, While trying to use rdd.take(numItems) My job just hangs there forever, the following is output messages: 14/03/07 00:52:21 INFO SparkContext: Starting job: take at xx.java:55 14/03/07 00:52:21 INFO DAGScheduler: Got job 1 (take at xx.java:55) with 1 output p