Yes I am all for it, as I use Hive with Oracle as its metastore extensively.
Case in point, on 6th March A Hive user <https://lists.apache.org/thread/vhgxt1cj2ppc862j0lwxl63j6nfc7khh> alluded to it and I quote "I just wanted to highlight that Hive 3.x line is EOL. It has various known security vulnerabilities, many serious bugs (including wrong results and data corruption), and lacks lots of improvements and major features that are available in Hive 4. Upgrading is the right path forward." In summary, Hive 4.x likely includes performance improvements, new features, and bug fixes. Compiling against it would allow Spark to take advantage of these. Plus using the latest versions of both Spark and Hive is important for maintaining a secure data platform. HTH Dr Mich Talebzadeh, Architect | Data Science | Financial Crime | Forensic Analysis | GDPR view my Linkedin profile <https://www.linkedin.com/in/mich-talebzadeh-ph-d-5205b2/> On Tue, 11 Mar 2025 at 19:08, Rozov, Vlad <vro...@amazon.com.invalid> wrote: > Hi All, > > As Apache Hive announced EOL for Hive 2.x [1] and 3.x [2], should Spark be > compiled against Hive 4.x and use it as default? > > Thank you, > > Vlad > > [1] https://lists.apache.org/thread/4ctrzfw60jkhc0hq2xoh1jpqxgt2zd93 > [2] https://lists.apache.org/thread/99h6wr7nk4684r6tkcbm8ydfytgqy6f3 > [3] https://github.com/apache/spark/pull/50213 >