I am working on it. Here is the (very rough) version: https://github.com/apache/spark/compare/apache:master...marmbrus:multiHiveVersions
On Mon, Apr 27, 2015 at 1:03 PM, Punyashloka Biswal <punya.bis...@gmail.com> wrote: > Thanks Marcelo and Patrick - I don't know how I missed that ticket in my > Jira search earlier. Is anybody working on the sub-issues yet, or is there > a design doc I should look at before taking a stab? > > Regards, > Punya > > On Mon, Apr 27, 2015 at 3:56 PM Patrick Wendell <pwend...@gmail.com> > wrote: > > > Hey Punya, > > > > There is some ongoing work to help make Hive upgrades more manageable > > and allow us to support multiple versions of Hive. Once we do that, it > > will be much easier for us to upgrade. > > > > https://issues.apache.org/jira/browse/SPARK-6906 > > > > - Patrick > > > > On Mon, Apr 27, 2015 at 12:47 PM, Marcelo Vanzin <van...@cloudera.com> > > wrote: > > > That's a lot more complicated than you might think. > > > > > > We've done some basic work to get HiveContext to compile against Hive > > > 1.1.0. Here's the code: > > > > > > https://github.com/cloudera/spark/commit/00e2c7e35d4ac236bcfbcd3d2805b483060255ec > > > > > > We didn't sent that upstream because that only solves half of the > > > problem; the hive-thriftserver is disabled in our CDH build because it > > > uses a lot of Hive APIs that have been removed in 1.1.0, so even > > > getting it to compile is really complicated. > > > > > > If there's interest in getting the HiveContext part fixed up I can > > > send a PR for that code. But at this time I don't really have plans to > > > look at the thrift server. > > > > > > > > > On Mon, Apr 27, 2015 at 11:58 AM, Punyashloka Biswal > > > <punya.bis...@gmail.com> wrote: > > >> Dear Spark devs, > > >> > > >> Is there a plan for staying up-to-date with current (and future) > > versions > > >> of Hive? Spark currently supports version 0.13 (June 2014), but the > > latest > > >> version of Hive is 1.1.0 (March 2015). I don't see any Jira tickets > > about > > >> updating beyond 0.13, so I was wondering if this was intentional or it > > was > > >> just that nobody had started work on this yet. > > >> > > >> I'd be happy to work on a PR for the upgrade if one of the core > > developers > > >> can tell me what pitfalls to watch out for. > > >> > > >> Punya > > > > > > > > > > > > -- > > > Marcelo > > > > > > --------------------------------------------------------------------- > > > To unsubscribe, e-mail: dev-unsubscr...@spark.apache.org > > > For additional commands, e-mail: dev-h...@spark.apache.org > > > > > >