Error happend at the "Lost task 0.0 in stage 0.0", I think it is not the
"groupBy" problem, it's the sql read the "customer" table issue,
please check the jdbc link and the data is loaded successfully??

Thanks
Xingchi

2016-01-11 15:43 GMT+08:00 Gaini Rajeshwar <raja.rajeshwar2...@gmail.com>:

> Hi All,
>
> I have a table named *customer *(customer_id, event, country, .... ) in
> postgreSQL database. This table is having more than 100 million rows.
>
> I want to know number of events from each country. To achieve that i am
> doing groupBY using spark as following.
>
> *val dataframe1 = sqlContext.load("jdbc", Map("url" ->
> "jdbc:postgresql://localhost/customerlogs?user=postgres&password=postgres",
> "dbtable" -> "customer"))*
>
>
> *dataframe1.groupBy("country").count().show()*
>
> above code seems to be getting complete customer table before doing
> groupBy. Because of that reason it is throwing the following error
>
> *16/01/11 12:49:04 WARN HeartbeatReceiver: Removing executor 0 with no
> recent heartbeats: 170758 ms exceeds timeout 120000 ms*
> *16/01/11 12:49:04 ERROR TaskSchedulerImpl: Lost executor 0 on 10.2.12.59
> <http://10.2.12.59>: Executor heartbeat timed out after 170758 ms*
> *16/01/11 12:49:04 WARN TaskSetManager: Lost task 0.0 in stage 0.0 (TID 0,
> 10.2.12.59): ExecutorLostFailure (executor 0 exited caused by one of the
> running tasks) Reason: Executor heartbeat timed out after 170758 ms*
>
> I am using spark 1.6.0
>
> Is there anyway i can solve this ?
>
> Thanks,
> Rajeshwar Gaini.
>

Reply via email to