hi! Are you compiling Spark itself? Do you use "./build/mvn" from the project root? If you compiled an other version of Spark before and there the scala version was different then zinc/nailgun could cached the old classes which can cause similar troubles. In that case this could help:
./build/zinc-0.3.15/bin/zinc -shutdown Best Regards, Attila On Wed, Mar 10, 2021 at 11:27 AM jiahong li <monkeyboy....@gmail.com> wrote: > hi, everybody, when i compile spark 3.1.1 from tag v3.1.1 ,encounter error > like this: > > INFO] --- scala-maven-plugin:4.3.0:compile (scala-compile-first) @ > spark-core_2.12 --- > [INFO] Using incremental compilation using Mixed compile order > [INFO] Compiler bridge file: > .sbt/1.0/zinc/org.scala-sbt/org.scala-sbt-compiler-bridge_2.12-1.3.1-bin_2.12.10__52.0-1.3.1_20191012T045515.jar > [INFO] compiler plugin: > BasicArtifact(com.github.ghik,silencer-plugin_2.12.10,1.6.0,null) > [INFO] Compiling 560 Scala sources and 99 Java sources to > git/spark/core/target/scala-2.12/classes ... > [ERROR] [Error] > git/spark/core/src/main/scala/org/apache/spark/ui/HttpSecurityFilter.scala:107: > type mismatch; > found : K where type K > required: String > [ERROR] [Error] > git/spark/core/src/main/scala/org/apache/spark/ui/HttpSecurityFilter.scala:107: > value map is not a member of V > [ERROR] [Error] > git/spark/core/src/main/scala/org/apache/spark/ui/HttpSecurityFilter.scala:107: > missing argument list for method stripXSS in class XssSafeRequest > Unapplied methods are only converted to functions when a function type is > expected. > You can make this conversion explicit by writing `stripXSS _` or > `stripXSS(_)` instead of `stripXSS`. > [ERROR] [Error] > git/spark/core/src/main/scala/org/apache/spark/ui/PagedTable.scala:307: > value startsWith is not a member of K > [ERROR] [Error] > git/spark/core/src/main/scala/org/apache/spark/util/Utils.scala:580: value > toLowerCase is not a member of object org.apache.hadoop.util.StringUtils > [ERROR] 5 errors found > > anybody encounter error like this? > >