It's good Thanks for your reply Michael. On Thu, Aug 20, 2015 at 11:03 PM, Michael Armbrust <mich...@databricks.com> wrote:
> We will probably fix this in Spark 1.6 > > https://issues.apache.org/jira/browse/SPARK-10040 > > On Thu, Aug 20, 2015 at 5:18 AM, Aram Mkrtchyan < > aram.mkrtchyan...@gmail.com> wrote: > >> We want to migrate our data (approximately 20M rows) from parquet to >> postgres, >> when we are using dataframe writer's jdbc method the execution time is very >> large, we have tried the same with batch insert it was much effective. >> Is it intentionally implemented in that way? >> > >