Hello Nicolas, Well the issue is that with Hive 3, Spark gets it's own metastore, separate from the Hive 3 metastore. So how do you reconcile this separation of metastores?
Can you continue to "enableHivemetastore" and be able to connect to Hive 3? Does this connection take advantage of Hive's LLAP? Our team doesn't believe that it's possible to make the connection as you would in the past. But if it is that simple, I would be ecstatic 😁. Thanks, Alfredo On Mon, Nov 18, 2019, 12:53 PM Nicolas Paris <nicolas.pa...@riseup.net> wrote: > Hi Alfredo > > my 2 cents: > To my knowlegde and reading the spark3 pre-release note, it will handle > hive metastore 2.3.5 - no mention of hive 3 metastore. I made several > tests on this in the past[1] and it seems to handle any hive metastore > version. > > However spark cannot read hive managed table AKA transactional tables. > So I would say you should be able to read any hive 3 regular table with > any of spark, pyspark or sparkR. > > > [1] > https://parisni.frama.io/posts/playing-with-hive-spark-metastore-versions/ > > On Mon, Nov 18, 2019 at 11:23:50AM -0600, Alfredo Marquez wrote: > > Hello, > > > > Our company is moving to Hive 3, and they are saying that there is no > SparkR > > implementation in Spark 2.3.x + that will connect to Hive 3. Is this > true? > > > > If it is true, will this be addressed in the Spark 3 release? > > > > I don't use python, so losing SparkR to get work done on Hadoop is a > huge loss. > > > > P.S. This is my first email to this community; if there is something I > should > > do differently, please let me know. > > > > Thank you > > > > Alfredo > > -- > nicolas > > --------------------------------------------------------------------- > To unsubscribe e-mail: user-unsubscr...@spark.apache.org > >