Spark SQL (which also can give you an RDD for use with the standard Spark RDD API) has support for json, parquet, and hive tables <http://spark.apache.org/docs/latest/sql-programming-guide.html#data-sources>. There is also a library for Avro <https://github.com/databricks/spark-avro>.
On Tue, Apr 14, 2015 at 2:49 AM, Akhil Das <ak...@sigmoidanalytics.com> wrote: > There's sc.objectFile also. > > Thanks > Best Regards > > On Tue, Apr 14, 2015 at 2:59 PM, ÐΞ€ρ@Ҝ (๏̯͡๏) <deepuj...@gmail.com> > wrote: > >> Can you please share the native support of data formats available with >> Spark. >> >> Two i can see are parquet and textFile >> >> sc.parquetFile >> sc.textFile >> >> I see that Hadoop Input Formats (Avro) are having issues, that i faced in >> earlier threads and seems to be well known. >> >> https://issues.apache.org/jira/browse/SPARK-993 >> >> https://issues.apache.org/jira/browse/SPARK-1018 >> >> >> Hence i want to know which data formats have full support in Spark. I can >> think of moving to those data formats. >> >> -- >> Deepak >> >> >