Spark SQL (which also can give you an RDD for use with the standard Spark
RDD API) has support for json, parquet, and hive tables
<http://spark.apache.org/docs/latest/sql-programming-guide.html#data-sources>.
There is also a library for Avro <https://github.com/databricks/spark-avro>.

On Tue, Apr 14, 2015 at 2:49 AM, Akhil Das <ak...@sigmoidanalytics.com>
wrote:

> There's sc.objectFile also.
>
> Thanks
> Best Regards
>
> On Tue, Apr 14, 2015 at 2:59 PM, ÐΞ€ρ@Ҝ (๏̯͡๏) <deepuj...@gmail.com>
> wrote:
>
>> Can you please share the native support of data formats available with
>> Spark.
>>
>> Two i can see are parquet and textFile
>>
>> sc.parquetFile
>> sc.textFile
>>
>> I see that Hadoop Input Formats (Avro) are having issues, that i faced in
>> earlier threads and seems to be well known.
>>
>> https://issues.apache.org/jira/browse/SPARK-993
>>
>> https://issues.apache.org/jira/browse/SPARK-1018
>>
>>
>> Hence i want to know which data formats have full support in Spark. I can
>> think of moving to those data formats.
>>
>> --
>> Deepak
>>
>>
>

Reply via email to