val conf = new SparkConf() .setAppName(detail)
.set("spark.serializer", "org.apache.spark.serializer.KryoSerializer") .set("spark.kryoserializer.buffer.mb", arguments.get("buffersize" ).get) .set("spark.kryoserializer.buffer.max.mb", arguments.get( "maxbuffersize").get) .set("spark.driver.maxResultSize", arguments.get("maxResultSize").get) .registerKryoClasses(Array(classOf[org.apache.accumulo.core.data.Key])) Can you try this ? On Tue, Apr 28, 2015 at 11:11 AM, madhvi <madhvi.gu...@orkash.com> wrote: > Hi, > > While connecting to accumulo through spark by making sparkRDD I am getting > the following error: > object not serializable (class: org.apache.accumulo.core.data.Key) > > This is due to the 'key' class of accumulo which does not implement > serializable interface.How it can be solved and accumulo can be used with > spark > > Thanks > Madhvi > > --------------------------------------------------------------------- > To unsubscribe, e-mail: user-unsubscr...@spark.apache.org > For additional commands, e-mail: user-h...@spark.apache.org > > -- Deepak