I see, I'm using Spark 1.6.0 and that change seems to be for 2.0 or maybe it's in 1.6.1 looking at the history. thanks I'll see if update spark to 1.6.1
On Thu, May 26, 2016 at 3:33 PM, Anthony May <anthony...@gmail.com> wrote: > It doesn't appear to be configurable, but it is inserting by column name: > > https://github.com/apache/spark/blob/master/sql/core/src/main/scala/org/apache/spark/sql/execution/datasources/jdbc/JdbcUtils.scala#L102 > > On Thu, 26 May 2016 at 16:02 Andrés Ivaldi <iaiva...@gmail.com> wrote: > >> Hello, >> I'realize that when dataframe executes insert it is inserting by scheme >> order column instead by name, ie >> >> dataframe.write(SaveMode).jdbc(url, table, properties) >> >> Reading the profiler the execution is >> >> insert into TableName values(a,b,c..) >> >> what i need is >> insert into TableNames (colA,colB,colC) values(a,b,c) >> >> could be some configuration? >> >> regards. >> >> -- >> Ing. Ivaldi Andres >> > -- Ing. Ivaldi Andres