Hi Sean, Thanks for bringing this up. Honestly, my opinion is that Spark should be fully ANSI SQL compliant. Where ANSI SQL compliance is not an issue, I am fine following any other DB. IMHO, we won't get anyway 100% compliance with any DB - postgres in this case (e.g. for decimal operations, we are following SQLServer, and postgres behaviour would be very hard to meet) - so I think it is fine that PMC members decide for each feature whether it is worth to support it or not.
Thanks, Marco On Mon, 8 Jul 2019, 20:09 Sean Owen, <sro...@apache.org> wrote: > See the particular issue / question at > https://github.com/apache/spark/pull/24872#issuecomment-509108532 and > the larger umbrella at > https://issues.apache.org/jira/browse/SPARK-27764 -- Dongjoon rightly > suggests this is a broader question. > > --------------------------------------------------------------------- > To unsubscribe e-mail: dev-unsubscr...@spark.apache.org > >