The Mahout 0.11.0 Shell requires Spark 1.3. Please try with Spark 1.3.1.
On 10/08/2015 10:37 PM, go canal wrote:
> I tried Spark 1.4.1, same error. Then I saw the same error from shell
> command. So I suspect that it is the environment configuration problem.
> I have followed this https://mahout.apache.org/general/downloads.html for
> Mahout configuration.
> So it seems to be a Spark configuration problem, I guess, although I can run
> spark-example without errors. Will need to figure out what are missing.
> thanks, canal
>
>
> On Monday, October 5, 2015 12:23 AM, Pat Ferrel
> <[email protected]> wrote:
>
>
> Mahout 0.11.0 is built on Spark 1.4 and so 1.5.1 is a bit unknown. I think
> the Mahout Shell does not run on 1.5.1.
>
> That may not be the error below, which is caused when Mahout tries to create
> a set of jars to use in the Spark executors. The code runs `mahout -spark
> classpath` to get these. So something is missing in your env in Eclipse. Does
> `mahout -spark classpath` run in a shell, if so check to see if you env
> matches in Eclipse.
>
> Also what are you trying to do? I have some example Spark Context creation
> code if you are using Mahout as a Library.
>
>
> On Oct 3, 2015, at 2:14 AM, go canal <[email protected]> wrote:
>
> Hello,I am running a very simple Mahout application in Eclipse, but got this
> error:
> Exception in thread "main" java.lang.IllegalArgumentException: Unable to read
> output from "mahout -spark classpath". Is SPARK_HOME defined?
> I have SPARK_HOME defined in Eclipse as an environment variable with value of
> /usr/local/spark-1.5.1.
> What else I need to include/set ?
>
> thanks, canal
>
>
>