Hi It worked for me like this. Just define the case class outside of any class to write to parquet format successfully. I am using Spark version 1.1.1.
case class person(id: Int, name: String, fathername: String, officeid: Int) object Program { def main (args: Array[String]) { val conf: SparkConf = new SparkConf().setAppName("Test").setMaster("local") val sc: SparkContext = new SparkContext(conf) val sqlContext = new org.apache.spark.sql.SQLContext(sc) import sqlContext.createSchemaRDD val baseperson: RDD[Array[String]] = sc.textFile("/home/ahsan/Desktop/1.csv").flatMap(line => line.split("\n")).map(_.split(",")) val x: RDD[person] = baseperson.map(p => person(p(0).asInstanceOf[Int], p(1).asInstanceOf[String], p(2).asInstanceOf[String], p(3).asInstanceOf[Int])) x.saveAsParquetFile("/home/ahsan/Desktop/pqt") } } -- View this message in context: http://apache-spark-user-list.1001560.n3.nabble.com/parquet-file-not-loading-spark-v-1-1-0-tp20618p20628.html Sent from the Apache Spark User List mailing list archive at Nabble.com. --------------------------------------------------------------------- To unsubscribe, e-mail: user-unsubscr...@spark.apache.org For additional commands, e-mail: user-h...@spark.apache.org