Hi, Michael, It works to me! Thanks a lot! If I use spark-hive or HiveContext, do I have to setup Hive on server? Can I run this on my local laptop?
On Thu, Sep 17, 2015 at 11:02 AM, Michael Armbrust <mich...@databricks.com> wrote: > libraryDependencies += "org.apache.spark" %% "spark-sql" % "1.4.1" > > Though, I would consider using spark-hive and HiveContext, as the query > parser is more powerful and you'll have access to window functions and other > features. > > > On Thu, Sep 17, 2015 at 10:59 AM, Cui Lin <icecreamlc...@gmail.com> wrote: > >> Hello, >> >> I got stuck in adding spark sql into my standalone application. >> The build.sbt is defined as: >> >> libraryDependencies += "org.apache.spark" %% "spark-core" % "1.4.1" >> >> >> I got the following error when building the package: >> >> *[error] /data/workspace/test/src/main/scala/TestMain.scala:6: object sql is >> not a member of package org.apache.spark >> [error] import org.apache.spark.sql.SQLContext; >> [error] ^ >> [error] /data/workspace/test/src/main/scala/TestMain.scala:19: object sql is >> not a member of package org.apache.spark >> [error] val sqlContext = new org.apache.spark.sql.SQLContext(sc) >> [error] ^ >> [error] two errors found >> [error] (compile:compile) Compilation failed* >> >> >> So sparksql is not part of spark core package? I have no issue when >> testing my codes in spark-shell. Thanks for the help! >> >> >> >> -- >> Best regards! >> >> Lin,Cui >> > > -- Best regards! Lin,Cui