Hi Will,

These three environment variables are needed [1].

I have had success with Hive 0.12 and Hadoop 1.0.4. For Hive, getting
the source distribution seems to be required. Docs contribution will
be much appreciated!

[1] 
https://github.com/apache/spark/tree/master/sql#other-dependencies-for-developers

Zongheng

On Thu, Jul 17, 2014 at 7:51 PM, Will Benton <wi...@redhat.com> wrote:
> Hi all,
>
> What's the preferred environment for generating golden test outputs for new 
> Hive tests?  In particular:
>
> * what Hadoop version and Hive version should I be using,
> * are there particular distributions people have run successfully, and
> * are there any system properties or environment variables (beyond 
> HADOOP_HOME, HIVE_HOME, and HIVE_DEV_HOME) I need to set before running the 
> suite?
>
> I ask because I'm getting some errors while trying to add new tests and would 
> like to eliminate any possible problems caused by differences between what my 
> environment offers and what Spark expects.  (I'm currently running with the 
> Fedora packages for Hadoop 2.2.0 and a locally-built Hive 0.12.0.)  Since 
> I'll only be using this for generating test outputs, something as simple to 
> set up as possible would be great.
>
> (Once I get something working, I'll be happy to write it up and contribute it 
> as developer docs.)
>
>
> thanks,
> wb

Reply via email to