Re: CSV Support in SparkR

2015-06-02 Thread Shivaram Venkataraman
quot;dev@spark.apache.org" < > dev@spark.apache.org> > > Subject: Re: CSV Support in SparkR > > Ah, alright, cool. I’ll rebuild and let you know. > > Thanks again, > Alek > > From: Shivaram Venkataraman > Reply-To: "shiva...@eecs.berkeley.edu" >

Re: CSV Support in SparkR

2015-06-02 Thread Eskilson,Aleksander
t;" mailto:shiva...@eecs.berkeley.edu>> Cc: Burak Yavuz mailto:brk...@gmail.com>>, "dev@spark.apache.org<mailto:dev@spark.apache.org>" mailto:dev@spark.apache.org>> Subject: Re: CSV Support in SparkR Ah, alright, cool. I’ll rebuild and let you know. Thank

Re: CSV Support in SparkR

2015-06-02 Thread Shivaram Venkataraman
showDF(df, 10) > > (my data is pipeline delimited, and the default SQL context is sqlCtx) > > Thanks, > Alek > > From: Shivaram Venkataraman > Reply-To: "shiva...@eecs.berkeley.edu" > Date: Tuesday, June 2, 2015 at 2:08 PM > To: Burak Yavuz > Cc

Re: CSV Support in SparkR

2015-06-02 Thread Eskilson,Aleksander
ailto:shiva...@eecs.berkeley.edu>" mailto:shiva...@eecs.berkeley.edu>> Date: Tuesday, June 2, 2015 at 2:08 PM To: Burak Yavuz mailto:brk...@gmail.com>> Cc: Aleksander Eskilson mailto:alek.eskil...@cerner.com>>, "dev@spark.apache.org<mailto:dev@spark.apache.org>&q

Re: CSV Support in SparkR

2015-06-02 Thread Eskilson,Aleksander
@cerner.com>>, "dev@spark.apache.org<mailto:dev@spark.apache.org>" mailto:dev@spark.apache.org>>, Shivaram Venkataraman mailto:shiva...@eecs.berkeley.edu>> Subject: Re: CSV Support in SparkR Hi Alek As Burak said, you can already use the spark-csv with SparkR in the

Re: CSV Support in SparkR

2015-06-02 Thread Shivaram Venkataraman
Hi Alek As Burak said, you can already use the spark-csv with SparkR in the 1.4 release. So right now I use it with something like this # Launch SparkR ./bin/sparkR --packages com.databricks:spark-csv_2.10:1.0.3 df <- read.df(sqlContext, "./nycflights13.csv", "com.databricks.spark.csv", header="t

Re: CSV Support in SparkR

2015-06-02 Thread Burak Yavuz
Hi, cc'ing Shivaram here, because he worked on this yesterday. If I'm not mistaken, you can use the following workflow: ```./bin/sparkR --packages com.databricks:spark-csv_2.10:1.0.3``` and then ```df <- read.df(sqlContext, "/data", "csv", header = "true")``` Best, Burak On Tue, Jun 2, 2015 a

CSV Support in SparkR

2015-06-02 Thread Eskilson,Aleksander
Are there any intentions to provide first class support for CSV files as one of the loadable file types in SparkR? Data brick’s spark-csv API [1] has support for SQL, Python, and Java/Scala, and implements most of the arguments of R’s read.table API [2], but currently there is no way to load CSV