Hi, hari,

I don't think job-server can work with SparkR (also pySpark). It seems it would 
be technically possible but needs support from job-server and SparkR(also 
pySpark), which doesn't exist yet.

But there may be some in-direct ways of sharing RDDs between SparkR and an 
application. For example, you may save the RDD in the application as a text 
file which can be later loaded by SparkR to create an SparkR RDD, or vice vesa. 
(unfortunately, SparkR now only supports creating RDDs from text files instead 
of any arbitrary Hadoop files.)  It would be easier if your application can use 
DataFrame. SparkR can create and manipulates DataFrames from virtually any 
external data sources, which can be used as intermediate media for data 
exchange.
________________________________________
From: harirajaram [hari.raja...@gmail.com]
Sent: Monday, July 13, 2015 8:30 PM
To: user@spark.apache.org
Subject: Share RDD from SparkR and another application

Hello,
I would like to share RDD between an application and sparkR.
I understand we have job-server and IBM kernel for sharing the context for
different applications but not sure how we can use it with sparkR as it is
some sort of front end (R shell) with spark.
Any insights appreciated.

Hari





--
View this message in context: 
http://apache-spark-user-list.1001560.n3.nabble.com/Share-RDD-from-SparkR-and-another-application-tp23795.html
Sent from the Apache Spark User List mailing list archive at Nabble.com.

---------------------------------------------------------------------
To unsubscribe, e-mail: user-unsubscr...@spark.apache.org
For additional commands, e-mail: user-h...@spark.apache.org


---------------------------------------------------------------------
To unsubscribe, e-mail: user-unsubscr...@spark.apache.org
For additional commands, e-mail: user-h...@spark.apache.org

Reply via email to