There is not an official updated version of Shark for Spark-1.0 (though you
might check out the untested spark-1.0 branch on the github).

You can also check out the preview release of Shark that runs on Spark SQL:
https://github.com/amplab/shark/tree/sparkSql

Michael


On Fri, Jun 6, 2014 at 6:02 AM, bijoy deb <bijoy.comput...@gmail.com> wrote:

> Hi,
>
> I am trying to run build Shark-0.9.1 from source,with Spark-1.0.0 as its
> dependency,using sbt package command.But I am getting the below error
> during build,which is making me think that perhaps Spark-1.0.0 is not
> compatible with Shark-0.9.1:
>
> [info]   Compilation completed in 9.046 s
>
>
>
>
>
>
>
>
>
>
>
> *[error] /vol1/shark/src/main/scala/shark/api/JavaTableRDD.scala:57:
> org.apache.spark.api.java.function.Function[shark.api.Row,Boolean] does not
> take parameters[error]     wrapRDD(rdd.filter((x => f(x).booleanValue())))
> [error]                               ^[error]
> /vol1/shark/src/main/scala/shark/execution/CoGroupedRDD.scala:84: type
> mismatch;[error]  found   : String[error]  required:
> org.apache.spark.serializer.Serializer [error]         new
> ShuffleDependency[Any, Any](rdd, part,
> SharkEnv.shuffleSerializerName)[error]
> ^[error] /vol1/shark/src/main/scala/shark/execution/CoGroupedRDD.scala:120:
> value serializerManager is not a member of org.apache.spark.SparkEnv
> [error]     val serializer =
> SparkEnv.get.serializerManager.get(SharkEnv.shuffleSerializerName,
> SparkEnv.get.conf)[error]                                   ^*[warn]
> /vol1/shark/src/main/scala/shark/execution/ExtractOperator.scala:111:
> non-variable type argument (shark.execution.ReduceKey, Any) in type pattern
> org.apache.spark.rdd.RDD[(shark.execution.ReduceKey, Any)] is unchecked
> since it is eliminated by erasure
> [warn]           case r: RDD[(ReduceKey, Any)] => RDDUtils.sortByKey(r)
> [warn]
>
>
>
>
> *                ^[error]
> /vol1/shark/src/main/scala/shark/execution/GroupByPostShuffleOperator.scala:204:
> type mismatch; [error]  found   : String[error]  required:
> org.apache.spark.serializer.Serializer[error]
> .setSerializer(SharkEnv.shuffleSerializerName)[error]   *
> ^
> .....
> ...
>
> Can you please suggest if there is any way to use the Shark with the new
> Spark-1.0.0 version?
>
> Thanks
> Bijoy
>

Reply via email to