There is a new thread here. https://lists.apache.org/thread.html/ra2418b75ac276861a598e7ec943750e2b038c2f8ba49f41db57e5ae9%40%3Cdev.spark.apache.org%3E
Could you share your use case of Hive 1.2 here, Koert? Bests, Dongjoon. On Wed, Oct 7, 2020 at 1:04 PM Koert Kuipers <ko...@tresata.com> wrote: > i am a little confused about this. i assumed spark would no longer make a > distribution with hive 1.x, but the hive-1.2 profile remains. > > yet i see the hive-1.2 profile has been removed from pom.xml? > > On Wed, Sep 23, 2020 at 6:58 PM Dongjoon Hyun <dongjoon.h...@gmail.com> > wrote: > >> Hi, All. >> >> Since Apache Spark 3.0.0, Apache Hive 2.3.7 is the default >> Hive execution library. The forked Hive 1.2.1 library is not >> recommended because it's not maintained properly. >> >> In Apache Spark 3.1 on December 2020, we are going to >> remove it from our official distribution. >> >> https://github.com/apache/spark/pull/29856 >> SPARK-32981 Remove hive-1.2/hadoop-2.7 from Apache Spark 3.1 >> distribution >> >> Of course, the users still can build it from the source because the >> profile `hive-1.2` is still alive. >> >> Please let us know if you are going to build with the forked unofficial >> Hive 1.2.1 library still in Apache Spark 3.1. We want to listen to your >> pain-points before moving forward in this area. Eventually we will remove >> Hive 1.2 as a last piece of migration to Hive 2.3/Hadoop3/Java11+. >> >> Bests, >> Dongjoon. >> >