Hi all,
I'm trying to use spark sql to store data in parquet file. I create the
file and insert data into it with the following code :
* val conf = new SparkConf().setAppName("MCT").setMaster("local[2]")
val sc = new SparkContext(conf) val sqlContext = new SQLContext(sc)
import sqlContext._ val personParquet =
createParquetFile[Person]("people_1.pqt")
personParquet.registerAsTable("people") val data =
sc.parallelize(Seq(Person("Toto", 10), Person("foo", 101)))
data.insertInto("people") personParquet.collect foreach(println)
data.insertInto("people") val personParquet2 =
parquetFile("people_1.pqt") personParquet2.collect foreach(println)*
It works as I expect when I run it in spark-shell. But with a stand alone
application, I get a build error :
*MCT.scala:18: not found: value createParquetFile*
If I skip this creation set and save the rdd as parquet file directly it
works. But then, when I insert new data nothing happen.
What I'm doing wrong ?
Best regards,
Jaonary