Requesting some suggestions on this.

Thanks in advance,
Sourabh Badhya

On Mon, May 9, 2022 at 5:16 PM Sourabh Badhya <sbad...@cloudera.com> wrote:

> Hi team,
>
> I would like to know the behaviour of Append & Overwrite modes when table
> is not present and whether automatic table creation is
> supported/unsupported when df.write is used in Spark 3 when the underlying
> custom datasource is using SupportCatalogOptions.
>
> As per my knowledge, in the current implementation in master,  df.write in
> Append and Overwrite mode tries to load the table and look for the schema
> of the table. Hence the table is expected to be created beforehand and
> hence the table will not automatically be created. Attaching the below code
> link for reference.
>
> Code link -
> https://github.com/apache/spark/blob/b065c945fe27dd5869b39bfeaad8e2b23a8835b5/sql/core/src/main/scala/org/apache/spark/sql/DataFrameWriter.scala#L287
>
> This is slightly different from the behaviour we observed in Spark 2 -
> https://lists.apache.org/thread/y468ngqhfhxhv0fygvwvy8r3g4sw9v7n
>
> Please confirm if I am correct and if this is the intended behaviour in
> Spark 3.
>
> Thanks and regards,
> Sourabh Badhya
>

Reply via email to