No, I am reading the data from hdfs, transforming it , registering the data in a temp table using registerTempTable and then doing insert overwrite using Spark SQl' hiveContext.
On Thu, Jun 9, 2016 at 3:40 PM, Mich Talebzadeh <mich.talebza...@gmail.com> wrote: > how are you doing the insert? from an existing table? > > Dr Mich Talebzadeh > > > > LinkedIn * > https://www.linkedin.com/profile/view?id=AAEAAAAWh2gBxianrbJd6zP6AcPCCdOABUrV8Pw > <https://www.linkedin.com/profile/view?id=AAEAAAAWh2gBxianrbJd6zP6AcPCCdOABUrV8Pw>* > > > > http://talebzadehmich.wordpress.com > > > > On 9 June 2016 at 21:16, Stephen Boesch <java...@gmail.com> wrote: > >> How many workers (/cpu cores) are assigned to this job? >> >> 2016-06-09 13:01 GMT-07:00 SRK <swethakasire...@gmail.com>: >> >>> Hi, >>> >>> How to insert data into 2000 partitions(directories) of ORC/parquet at a >>> time using Spark SQL? It seems to be not performant when I try to insert >>> 2000 directories of Parquet/ORC using Spark SQL. Did anyone face this >>> issue? >>> >>> Thanks! >>> >>> >>> >>> -- >>> View this message in context: >>> http://apache-spark-user-list.1001560.n3.nabble.com/How-to-insert-data-into-2000-partitions-directories-of-ORC-parquet-at-a-time-using-Spark-SQL-tp27132.html >>> Sent from the Apache Spark User List mailing list archive at Nabble.com. >>> >>> --------------------------------------------------------------------- >>> To unsubscribe, e-mail: user-unsubscr...@spark.apache.org >>> For additional commands, e-mail: user-h...@spark.apache.org >>> >>> >> >