I think we can expect Spark 2.0 support to be added in one of the nearest releases (not sure if it will be 1.8 or 2.0).
On Mon, Sep 19, 2016 at 10:20 PM, Richard Siebeling <rsiebel...@gmail.com> wrote: > I'd wish someone could fix this, unfortunately I can't do anything about it > in the near future... > > On Mon, Sep 19, 2016 at 3:20 PM, Phadnis, Varun <phad...@sky.optymyze.com> > wrote: > > > Hello, > > > > Just to be clear, I am not explicitly building the Hadoop edition. I get > > the described error when I simply specify the Spark version: > > > > mvn clean package -Dspark.version=2.0.0 -DskipTests > > > > Thanks for the quick response! > > > > > > -----Original Message----- > > From: Sergey Kozlov [mailto:skoz...@gridgain.com] > > Sent: 19 September 2016 05:56 > > To: dev@ignite.apache.org > > Subject: Re: Support for Spark 2.0 > > > > Hi > > > > It's a known issue IGNITE-3596 Hadoop edition can't be compiled against > > spark 2.0.0 <https://issues.apache.org/jira/browse/IGNITE-3596> > > > > Unfortunately there's no progress yet > > > > On Mon, Sep 19, 2016 at 1:36 PM, Phadnis, Varun < > phad...@sky.optymyze.com> > > wrote: > > > > > Hello, > > > > > > Can someone please tell me when is the support for Spark 2.0 planned? > > > > > > Currently Ignite cannot be built for Spark 2.0. Attempting this yields > > > the following error : > > > > > > > > > ---------------------------------------------------------------------- > > > -- > > > > > > [INFO] Building ignite-spark 1.8.0-SNAPSHOT [INFO] > > > > > > ---------------------------------------------------------------------- > > > -- > > > > > > [INFO] > > > > > > [INFO] --- maven-clean-plugin:2.5:clean (default-clean) @ ignite-spark > > > --- [INFO] Deleting /home/spark/code/ignite/modules/spark/target > > > > > > [INFO] > > > > > > [INFO] --- flatten-maven-plugin:1.0.0-beta-3:clean > > > (flatten.clean.before) @ ignite-spark --- [INFO] Deleting > > > /home/spark/code/ignite/ modules/spark/pom-installed.xml > > > > > > [INFO] > > > > > > [INFO] --- maven-enforcer-plugin:1.4:enforce (default) @ ignite-spark > > > --- [INFO] [INFO] --- maven-remote-resources-plugin:1.5:process > > > (default) @ ignite-spark --- [INFO] [INFO] --- > > > maven-remote-resources-plugin:1.5:process > > > (ignite-dependencies) @ ignite-spark --- [INFO] [INFO] --- > > > maven-resources-plugin:2.6:resources (default-resources) @ > > > ignite-spark > > > --- [INFO] Using 'UTF-8' encoding to copy filtered resources. > > > > > > [INFO] skip non existing resourceDirectory /home/spark/code/ignite/ > > > modules/spark/src/main/resources > > > > > > [INFO] Copying 4 resources > > > > > > [INFO] Copying 4 resources > > > > > > [INFO] > > > > > > [INFO] --- flatten-maven-plugin:1.0.0-beta-3:flatten (flatten) @ > > > ignite-spark --- [INFO] Generating flattened POM of project > > > org.apache.ignite:ignite-spark:jar:1.8.0-SNAPSHOT... > > > > > > [WARNING] Ignoring multiple XML header comment! > > > > > > [INFO] > > > > > > [INFO] --- scala-maven-plugin:3.2.0:add-source (scala-compile-first) @ > > > ignite-spark --- [INFO] Add Source directory: > > > > > > /home/spark/code/ignite/modules/spark/src/main/scala > > > > > > [INFO] Add Test Source directory: > > > > > > /home/spark/code/ignite/modules/spark/src/test/scala > > > > > > [INFO] > > > > > > [INFO] --- scala-maven-plugin:3.2.0:compile (scala-compile-first) @ > > > ignite-spark --- [WARNING] Expected all dependencies to require Scala > > > version: 2.11.7 [WARNING] > > > org.apache.ignite:ignite-spark:1.8.0-SNAPSHOT > > > requires scala > > > > > > version: 2.11.7 > > > > > > [WARNING] com.twitter:chill_2.11:0.8.0 requires scala version: 2.11.7 > > > [WARNING] org.apache.spark:spark-core_2.11:2.0.0 requires scala > > version: > > > > > > 2.11.8 > > > > > > [WARNING] Multiple versions of scala libraries detected! > > > > > > [INFO] /home/spark/code/ignite/modules/spark/src/main/scala:-1: info: > > > > > > compiling > > > > > > [INFO] Compiling 8 source files to > > > > > > /home/spark/code/ignite/modules/spark/target/classes at 1474028222935 > > > [ERROR] > > > > > > /home/spark/code/ignite/modules/spark/src/main/scala/ > > > org/apache/ignite/spark/IgniteContext.scala:25: > > > > > > error: object Logging is not a member of package org.apache.spark > > > [ERROR] import org.apache.spark.{Logging, SparkContext} > > > > > > [ERROR] ^ > > > > > > [ERROR] > > > > > > /home/spark/code/ignite/modules/spark/src/main/scala/ > > > org/apache/ignite/spark/IgniteContext.scala:37: > > > > > > error: not found: type Logging > > > > > > [ERROR] ) extends Serializable with Logging { > > > > > > [ERROR] ^ > > > > > > [ERROR] > > > > > > /home/spark/code/ignite/modules/spark/src/main/scala/ > > > org/apache/ignite/spark/IgniteContext.scala:50: > > > > > > error: not found: value logInfo > > > > > > [ERROR] logInfo("Will start Ignite nodes on " + workers + " > > > > > > workers") > > > > > > [ERROR] ^ > > > > > > [ERROR] > > > > > > /home/spark/code/ignite/modules/spark/src/main/scala/ > > > org/apache/ignite/spark/IgniteContext.scala:129: > > > > > > error: not found: value logInfo > > > > > > [ERROR] logInfo("Setting IGNITE_HOME from driver not as it > is > > > > > > not available on this worker: " + igniteHome) > > > > > > [ERROR] ^ > > > > > > [ERROR] > > > > > > /home/spark/code/ignite/modules/spark/src/main/scala/ > > > org/apache/ignite/spark/IgniteContext.scala:146: > > > > > > error: not found: value logError > > > > > > [ERROR] logError("Failed to start Ignite.", e) > > > > > > [INFO] ^ > > > > > > [ERROR] > > > > > > /home/spark/code/ignite/modules/spark/src/main/scala/ > > > org/apache/ignite/spark/IgniteContext.scala:164: > > > > > > error: not found: value logInfo > > > > > > [ERROR] logInfo("Will stop Ignite nodes on " + workers > + > > " > > > > > > workers") > > > > > > [ERROR] ^ > > > > > > [ERROR] 6 errors found > > > > > > [INFO] > > > > > > ---------------------------------------------------------------------- > > > -- > > > > > > Thanks! > > > > > > > > > > > -- > > Sergey Kozlov > > GridGain Systems > > www.gridgain.com > > >