i use ./build/mvn to compile ,and after execute command
:./build/zinc-0.3.15/bin/zinc
-shutdown
and execute command like this: /dev/make-distribution.sh --name
custom-spark --pip  --tgz -Phive -Phive-thriftserver -Pyarn
-Dhadoop.version=2.6.0-cdh5.13.1 -DskipTests
same error appear.
and execute command: ps -ef |grep zinc, there is nothing containe zinc

Attila Zsolt Piros <piros.attila.zs...@gmail.com> 于2021年3月10日周三 下午6:55写道:

> hi!
>
> Are you compiling Spark itself?
> Do you use "./build/mvn" from the project root?
> If you compiled an other version of Spark before and there the scala
> version was different then zinc/nailgun could cached the old classes which
> can cause similar troubles.
> In that case this could help:
>
> ./build/zinc-0.3.15/bin/zinc -shutdown
>
> Best Regards,
> Attila
>
> On Wed, Mar 10, 2021 at 11:27 AM jiahong li <monkeyboy....@gmail.com>
> wrote:
>
>> hi, everybody, when i compile spark 3.1.1 from tag v3.1.1 ,encounter
>> error like this:
>>
>> INFO] --- scala-maven-plugin:4.3.0:compile (scala-compile-first) @
>> spark-core_2.12 ---
>> [INFO] Using incremental compilation using Mixed compile order
>> [INFO] Compiler bridge file:
>> .sbt/1.0/zinc/org.scala-sbt/org.scala-sbt-compiler-bridge_2.12-1.3.1-bin_2.12.10__52.0-1.3.1_20191012T045515.jar
>> [INFO] compiler plugin:
>> BasicArtifact(com.github.ghik,silencer-plugin_2.12.10,1.6.0,null)
>> [INFO] Compiling 560 Scala sources and 99 Java sources to
>> git/spark/core/target/scala-2.12/classes ...
>> [ERROR] [Error]
>> git/spark/core/src/main/scala/org/apache/spark/ui/HttpSecurityFilter.scala:107:
>> type mismatch;
>>  found   : K where type K
>>  required: String
>> [ERROR] [Error]
>> git/spark/core/src/main/scala/org/apache/spark/ui/HttpSecurityFilter.scala:107:
>> value map is not a member of V
>> [ERROR] [Error]
>> git/spark/core/src/main/scala/org/apache/spark/ui/HttpSecurityFilter.scala:107:
>> missing argument list for method stripXSS in class XssSafeRequest
>> Unapplied methods are only converted to functions when a function type is
>> expected.
>> You can make this conversion explicit by writing `stripXSS _` or
>> `stripXSS(_)` instead of `stripXSS`.
>> [ERROR] [Error]
>> git/spark/core/src/main/scala/org/apache/spark/ui/PagedTable.scala:307:
>> value startsWith is not a member of K
>> [ERROR] [Error]
>> git/spark/core/src/main/scala/org/apache/spark/util/Utils.scala:580: value
>> toLowerCase is not a member of object org.apache.hadoop.util.StringUtils
>> [ERROR] 5 errors found
>>
>> anybody encounter error like this?
>>
>>
>

Reply via email to