Hi,
Is there any recommended way of serializing Hadoop Writables' in Spark?
Here is my problem.
Question1:
I have a pair RDD which is created by reading a SEQ[LongWritable,
BytesWritable]:
RDD[(LongWritable, BytesWritable)]
I have these two settings set in spark conf.
spark.serializer=org.apach
Hi,
Is there any recommended way of serializing Hadoop Writables' in Spark?
Here is my problem.
Question1:
I have a pair RDD which is created by reading a SEQ[LongWritable,
BytesWritable]:
RDD[(LongWritable, BytesWritable)]
I have these two settings set in spark conf.
spark.serializer=org.apach
Hello Everyone,
I have my parquet files stored on HDFS. I am trying to create a table in
Hive Metastore from Spark SQL. I have an Avro schema file from which I
generated the parquet files.
I am doing the following to create the table.
1) Firstly create an Avro dummy table from the schema file