Just add spark_1.4.1_yarn_shuffle.jar in ClassPath or create a New Maven project using below dependency:
<dependency> <groupId>org.apache.spark</groupId> <artifactId>spark-core_2.11</artifactId> <version>1.4.1</version> </dependency> <dependency> <groupId>org.apache.spark</groupId> <artifactId>spark-sql_2.11</artifactId> <version>1.4.1</version> </dependency> On Tue, Aug 18, 2015 at 11:51 PM, Jerry <jerry.c...@gmail.com> wrote: > So from what I understand, those usually pull dependencies for a given > project? I'm able to run the spark shell so I'd assume I have everything. > What am I missing from the big picture and what directory do I run maven on? > > Thanks, > Jerry > > On Tue, Aug 18, 2015 at 11:15 AM, Ted Yu <yuzhih...@gmail.com> wrote: > >> Normally people would establish maven project with Spark dependencies or, >> use sbt. >> >> Can you go with either approach ? >> >> Cheers >> >> On Tue, Aug 18, 2015 at 10:28 AM, Jerry <jerry.c...@gmail.com> wrote: >> >>> Hello, >>> >>> So I setup Spark to run on my local machine to see if I can reproduce >>> the issue I'm having with data frames, but I'm running into issues with the >>> compiler. >>> >>> Here's what I got: >>> >>> $ echo $CLASSPATH >>> >>> /usr/lib/jvm/java-6-oracle/lib:/home/adminz/dev/spark/spark-1.4.1/lib/spark-assembly-1.4.1-hadoop2.6.0.jar >>> >>> >>> javac Test.java >>> Test.java:1: package org.apache.spark.sql.api.java does not exist >>> import org.apache.spark.sql.api.java.*; >>> ^ >>> Test.java:6: package org.apache.spark.sql does not exist >>> import org.apache.spark.sql.*; >>> ^ >>> Test.java:7: package org.apache.spark.sql.hive does not exist >>> import org.apache.spark.sql.hive.*; >>> .... >>> >>> >>> Let me know what I'm doing wrong. >>> >>> Thanks, >>> Jerry >>> >> >> >