You could try following this guidelines http://docs.sigmoidanalytics.com/index.php/How_to_build_SPARK_on_Windows
Thanks Best Regards On Wed, Nov 26, 2014 at 12:24 PM, Sunita Arvind <sunitarv...@gmail.com> wrote: > Hi All, > > I just installed a spark on my laptop and trying to get spark-shell to > work. Here is the error I see: > > C:\spark\bin>spark-shell > Exception in thread "main" java.util.NoSuchElementException: key not > found: CLAS > SPATH > at scala.collection.MapLike$class.default(MapLike.scala:228) > at scala.collection.AbstractMap.default(Map.scala:58) > at scala.collection.MapLike$class.apply(MapLike.scala:141) > at scala.collection.AbstractMap.apply(Map.scala:58) > at > org.apache.spark.deploy.SparkSubmitDriverBootstrapper$.main(SparkSubm > itDriverBootstrapper.scala:49) > at > org.apache.spark.deploy.SparkSubmitDriverBootstrapper.main(SparkSubmi > tDriverBootstrapper.scala) > > > The classpath seems to be right: > > C:\spark\bin>compute-classpath.cmd > > ;;C:\spark\bin\..\conf;C:\spark\bin\..\lib\spark-assembly-1.1.0-hadoop2.3.0.jar; > > ;C:\spark\bin\..\lib\datanucleus-api-jdo-3.2.1.jar;C:\spark\bin\..\lib\datanucle > us-core-3.2.2.jar;C:\spark\bin\..\lib\datanucleus-rdbms-3.2.1.jar > > Manually exporting the classpath to include the assembly jar doesnt help > either. > > What could be wrong with this installation? Scala and SBT are installed, > in path and are working fine. > > Appreciate your help. > regards > Sunita > > >