All of these works

Thanks
-Mo


2014-04-09 2:34 GMT-04:00 Xiangrui Meng <[email protected]>:

> After sbt/sbt gen-diea, do not import as an SBT project but choose
> "open project" and point it to the spark folder. -Xiangrui
>
> On Tue, Apr 8, 2014 at 10:45 PM, Sean Owen <[email protected]> wrote:
> > I let IntelliJ read the Maven build directly and that works fine.
> > --
> > Sean Owen | Director, Data Science | London
> >
> >
> > On Wed, Apr 9, 2014 at 6:14 AM, Dong Mo <[email protected]> wrote:
> >> Dear list,
> >>
> >> SBT compiles fine, but when I do the following:
> >> sbt/sbt gen-idea
> >> import project as SBT project to IDEA 13.1
> >> Make Project
> >> and these errors show up:
> >>
> >> Error:(28, 8) object FileContext is not a member of package
> >> org.apache.hadoop.fs
> >> import org.apache.hadoop.fs.{FileContext, FileStatus, FileSystem, Path,
> >> FileUtil}
> >>        ^
> >> Error:(31, 8) object Master is not a member of package
> >> org.apache.hadoop.mapred
> >> import org.apache.hadoop.mapred.Master
> >>        ^
> >> Error:(34, 26) object yarn is not a member of package org.apache.hadoop
> >> import org.apache.hadoop.yarn.api._
> >>                          ^
> >> Error:(35, 26) object yarn is not a member of package org.apache.hadoop
> >> import org.apache.hadoop.yarn.api.ApplicationConstants.Environment
> >>                          ^
> >> Error:(36, 26) object yarn is not a member of package org.apache.hadoop
> >> import org.apache.hadoop.yarn.api.protocolrecords._
> >>                          ^
> >> Error:(37, 26) object yarn is not a member of package org.apache.hadoop
> >> import org.apache.hadoop.yarn.api.records._
> >>                          ^
> >> Error:(38, 26) object yarn is not a member of package org.apache.hadoop
> >> import org.apache.hadoop.yarn.client.YarnClientImpl
> >>                          ^
> >> Error:(39, 26) object yarn is not a member of package org.apache.hadoop
> >> import org.apache.hadoop.yarn.conf.YarnConfiguration
> >>                          ^
> >> Error:(40, 26) object yarn is not a member of package org.apache.hadoop
> >> import org.apache.hadoop.yarn.ipc.YarnRPC
> >>                          ^
> >> Error:(41, 26) object yarn is not a member of package org.apache.hadoop
> >> import org.apache.hadoop.yarn.util.{Apps, Records}
> >>                          ^
> >> Error:(49, 11) not found: type YarnClientImpl
> >>   extends YarnClientImpl with Logging {
> >>           ^
> >> Error:(48, 20) not found: type ClientArguments
> >> class Client(args: ClientArguments, conf: Configuration, sparkConf:
> >> SparkConf)
> >>                    ^
> >> Error:(51, 18) not found: type ClientArguments
> >>   def this(args: ClientArguments, sparkConf: SparkConf) =
> >>                  ^
> >> Error:(54, 18) not found: type ClientArguments
> >>   def this(args: ClientArguments) = this(args, new SparkConf())
> >>                  ^
> >> Error:(56, 12) not found: type YarnRPC
> >>   var rpc: YarnRPC = YarnRPC.create(conf)
> >>            ^
> >> Error:(56, 22) not found: value YarnRPC
> >>   var rpc: YarnRPC = YarnRPC.create(conf)
> >>                      ^
> >> Error:(57, 17) not found: type YarnConfiguration
> >>   val yarnConf: YarnConfiguration = new YarnConfiguration(conf)
> >>                 ^
> >> Error:(57, 41) not found: type YarnConfiguration
> >>   val yarnConf: YarnConfiguration = new YarnConfiguration(conf)
> >>                                         ^
> >> Error:(58, 59) value getCredentials is not a member of
> >> org.apache.hadoop.security.UserGroupInformation
> >>   val credentials =
> UserGroupInformation.getCurrentUser().getCredentials()
> >>                                                           ^
> >> Error:(60, 34) not found: type ClientDistributedCacheManager
> >>   private val distCacheMgr = new ClientDistributedCacheManager()
> >>                                  ^
> >> Error:(72, 5) not found: value init
> >>     init(yarnConf)
> >>     ^
> >> Error:(73, 5) not found: value start
> >>     start()
> >>     ^
> >> Error:(76, 24) value getNewApplication is not a member of
> >> org.apache.spark.Logging
> >>     val newApp = super.getNewApplication()
> >>                        ^
> >> Error:(137, 35) not found: type GetNewApplicationResponse
> >>   def verifyClusterResources(app: GetNewApplicationResponse) = {
> >>                                   ^
> >> Error:(156, 65) not found: type ApplicationSubmissionContext
> >>   def createApplicationSubmissionContext(appId: ApplicationId):
> >> ApplicationSubmissionContext = {
> >>                                                                 ^
> >> Error:(156, 49) not found: type ApplicationId
> >>   def createApplicationSubmissionContext(appId: ApplicationId):
> >> ApplicationSubmissionContext = {
> >>                                                 ^
> >> Error:(118, 31) not found: type ApplicationId
> >>   def getAppStagingDir(appId: ApplicationId): String = {
> >>                               ^
> >> Error:(224, 69) not found: type LocalResource
> >>   def prepareLocalResources(appStagingDir: String): HashMap[String,
> >> LocalResource] = {
> >>                                                                     ^
> >> Error:(307, 39) not found: type LocalResource
> >>       localResources: HashMap[String, LocalResource],
> >>                                       ^
> >> Error:(343, 38) not found: type ContainerLaunchContext
> >>       env: HashMap[String, String]): ContainerLaunchContext = {
> >>                                      ^
> >> Error:(341, 15) not found: type GetNewApplicationResponse
> >>       newApp: GetNewApplicationResponse,
> >>               ^
> >> Error:(342, 39) not found: type LocalResource
> >>       localResources: HashMap[String, LocalResource],
> >>                                       ^
> >> Error:(426, 11) value submitApplication is not a member of
> >> org.apache.spark.Logging
> >>     super.submitApplication(appContext)
> >>           ^
> >> Error:(423, 29) not found: type ApplicationSubmissionContext
> >>   def submitApp(appContext: ApplicationSubmissionContext) = {
> >>                             ^
> >> Error:(429, 33) not found: type ApplicationId
> >>   def monitorApplication(appId: ApplicationId): Boolean = {
> >>                                 ^
> >> Error:(123, 25) not found: type YarnClusterMetrics
> >>     val clusterMetrics: YarnClusterMetrics = super.getYarnClusterMetrics
> >>                         ^
> >> Error:(123, 52) value getYarnClusterMetrics is not a member of
> >> org.apache.spark.Logging
> >>     val clusterMetrics: YarnClusterMetrics = super.getYarnClusterMetrics
> >>                                                    ^
> >> Error:(127, 20) not found: type QueueInfo
> >>     val queueInfo: QueueInfo = super.getQueueInfo(args.amQueue)
> >>                    ^
> >> Error:(127, 38) value getQueueInfo is not a member of
> >> org.apache.spark.Logging
> >>     val queueInfo: QueueInfo = super.getQueueInfo(args.amQueue)
> >>                                      ^
> >> Error:(158, 22) not found: value Records
> >>     val appContext =
> >> Records.newRecord(classOf[ApplicationSubmissionContext])
> >>                      ^
> >> Error:(219, 14) not found: value FileContext
> >>     val fc = FileContext.getFileContext(qualPath.toUri(), conf)
> >>              ^
> >> Error:(230, 29) not found: value Master
> >>     val delegTokenRenewer = Master.getMasterPrincipal(conf)
> >>                             ^
> >> Error:(242, 13) value addDelegationTokens is not a member of
> >> org.apache.hadoop.fs.FileSystem
> >>       dstFs.addDelegationTokens(delegTokenRenewer, credentials)
> >>             ^
> >> Error:(244, 42) not found: type LocalResource
> >>     val localResources = HashMap[String, LocalResource]()
> >>                                          ^
> >> Error:(302, 43) value addCredentials is not a member of
> >> org.apache.hadoop.security.UserGroupInformation
> >>     UserGroupInformation.getCurrentUser().addCredentials(credentials)
> >>                                           ^
> >> Error:(323, 5) not found: value Apps
> >>     Apps.setEnvFromInputString(env,
> System.getenv("SPARK_YARN_USER_ENV"))
> >>     ^
> >> Error:(330, 36) not found: type ClientArguments
> >>   def userArgsToString(clientArgs: ClientArguments): String = {
> >>                                    ^
> >> Error:(345, 23) not found: value Records
> >>     val amContainer = Records.newRecord(classOf[ContainerLaunchContext])
> >>                       ^
> >> Error:(363, 16) not found: value Environment
> >>       new Path(Environment.PWD.$(),
> >> YarnConfiguration.DEFAULT_CONTAINER_TEMP_DIR) + " "
> >>                ^
> >> Error:(392, 21) not found: value Environment
> >>       javaCommand = Environment.JAVA_HOME.$() + "/bin/java"
> >>                     ^
> >> Error:(405, 16) not found: value ApplicationConstants
> >>       " 1> " + ApplicationConstants.LOG_DIR_EXPANSION_VAR + "/stdout" +
> >>                ^
> >> Error:(410, 22) not found: value Records
> >>     val capability =
> >> Records.newRecord(classOf[Resource]).asInstanceOf[Resource]
> >>                      ^
> >> Error:(410, 72) not found: type Resource
> >>     val capability =
> >> Records.newRecord(classOf[Resource]).asInstanceOf[Resource]
> >>                                                                        ^
> >> Error:(434, 26) value getApplicationReport is not a member of
> >> org.apache.spark.Logging
> >>       val report = super.getApplicationReport(appId)
> >>                          ^
> >> Error:(474, 20) not found: type ClientArguments
> >>     val args = new ClientArguments(argStrings, sparkConf)
> >>                    ^
> >> Error:(481, 31) not found: value YarnConfiguration
> >>     for (c <-
> conf.getStrings(YarnConfiguration.YARN_APPLICATION_CLASSPATH))
> >> {
> >>                               ^
> >> Error:(487, 5) not found: value Apps
> >>     Apps.addToEnvironment(env, Environment.CLASSPATH.name,
> >> Environment.PWD.$())
> >>     ^
> >> Error:(490, 7) not found: value Apps
> >>       Apps.addToEnvironment(env, Environment.CLASSPATH.name,
> >> Environment.PWD.$() +
> >>       ^
> >> Error:(496, 7) not found: value Apps
> >>       Apps.addToEnvironment(env, Environment.CLASSPATH.name,
> >> Environment.PWD.$() +
> >>       ^
> >> Error:(499, 5) not found: value Apps
> >>     Apps.addToEnvironment(env, Environment.CLASSPATH.name,
> >> Environment.PWD.$() +
> >>     ^
> >> Error:(504, 7) not found: value Apps
> >>       Apps.addToEnvironment(env, Environment.CLASSPATH.name,
> >> Environment.PWD.$() +
> >>       ^
> >> Error:(507, 5) not found: value Apps
> >>     Apps.addToEnvironment(env, Environment.CLASSPATH.name,
> >> Environment.PWD.$() +
> >>     ^
> >>
> >> Any idea what's causing them, and maybe I am not using the best
> practice to
> >> import Spark to IDE?
> >>
> >> I would appreciate any suggestion on the best practice to import Spark
> to
> >> any IDE.
> >>
> >> Thank you
> >>
> >> -Mo
>

Reply via email to