Hi all, I might be missing something, but does the new Spark 1.3 sqlContext save interface support using Avro as the schema structure when writing Parquet files, in a similar way to AvroParquetWriter (which I've got working)?
I've seen how you can load an avro file and save it as parquet from https://databricks.com/blog/2015/03/24/spark-sql-graduates-from-alpha-in-spark-1-3.html, but not using the 2 together. Thanks, and apologies if I've missed something obvious! Ewan