Any example code? In pyspark: sqlContex.sql("use mytable") my_df.saveAsTable("tmp_spark_debug", mode="overwrite")
1. The code above seems not register the table in hive. I have to create table from hdfs in hive, it reports some format error: rcformat and parquet. 2. Rerun the saveAsTable using mode="overwrite" in saveAsTable, it reports the table already exists. 3. Sometimes it creates a directory in hive/warehouse/tmp_spark_debug, not in hive/warehouse/mytable/tmp_spark_debug. My goal is simple: df.saveAsTable('blablabla') // create a hive table in some database, then it can be visited by hive. I tried lots of time, it seems there are lots of bug in pyspark. Or my mehtod is wrong? 2016-02-21 10:04 GMT+08:00 gen tang <gen.tan...@gmail.com>: > Hi, > > You can use > sqlContext.sql("use <your database>") > before use dataframe.saveAsTable > > Hope it could be helpful > > Cheers > Gen > > > On Sun, Feb 21, 2016 at 9:55 AM, Glen <cng...@gmail.com> wrote: > >> For dataframe in spark, so the table can be visited by hive. >> >> -- >> Jacky Wang >> > > -- Jacky Wang