Thanks Das and Ayan.
Do you have any refrences on how to create connection pool for hbase inside
foreachpartitions as mentioned in guide. In my case, I have to use kerberos
hbase cluster.
On Wed, Sep 21, 2016 at 6:39 PM, Tathagata Das
wrote:
> http://spark.apache.org/docs/latest/streaming-progr
http://spark.apache.org/docs/latest/streaming-programming-guide.html#design-patterns-for-using-foreachrdd
On Wed, Sep 21, 2016 at 4:26 PM, ayan guha wrote:
> Connection object is not serialisable. You need to implement a getorcreate
> function which would run on each executors to create hbase co
Connection object is not serialisable. You need to implement a getorcreate
function which would run on each executors to create hbase connection
locally.
On 22 Sep 2016 08:34, "KhajaAsmath Mohammed"
wrote:
> Hello Everyone,
>
> I am running spark application to push data from kafka. I am able to