I'd prefer simply not making Hadoop 3 the default until 3.1+, rather than introduce yet another build combination. Does Hadoop 2 + Hive 2 work and is there demand for it?
On Sat, Nov 16, 2019 at 3:52 AM Wenchen Fan <cloud0...@gmail.com> wrote: > > Do we have a limitation on the number of pre-built distributions? Seems this > time we need > 1. hadoop 2.7 + hive 1.2 > 2. hadoop 2.7 + hive 2.3 > 3. hadoop 3 + hive 2.3 > > AFAIK we always built with JDK 8 (but make it JDK 11 compatible), so don't > need to add JDK version to the combination. > > On Sat, Nov 16, 2019 at 4:05 PM Dongjoon Hyun <dongjoon.h...@gmail.com> wrote: >> >> Thank you for suggestion. >> >> Having `hive-2.3` profile sounds good to me because it's orthogonal to >> Hadoop 3. >> IIRC, originally, it was proposed in that way, but we put it under >> `hadoop-3.2` to avoid adding new profiles at that time. >> >> And, I'm wondering if you are considering additional pre-built distribution >> and Jenkins jobs. >> >> Bests, >> Dongjoon. >> --------------------------------------------------------------------- To unsubscribe e-mail: dev-unsubscr...@spark.apache.org