That's right, Spark needs Maven 3.6.0. Just install it locally and then configure IntelliJ to use the local Maven, not built-in 3.3.9. The docs you're looking at are actually in github.com/apache/spark-website as they aren't version-specific and tied to a release.
On Sun, Apr 14, 2019 at 12:00 PM William Wong <william1...@gmail.com> wrote: > > Hi Sean, > > I would like to open a PR for updating the documentation. However, i cannot > share any file for 'http://spark.apache.org/developer-tools.html' It seems > that this file is not a part of the documentation (under folder docs)... > > Thanks and regards, > William > > On Sun, Apr 14, 2019 at 11:58 PM William Wong <william1...@gmail.com> wrote: >> >> Hi Sean, >> >> I tried the button, but antlr4 source was not generate as expected. I >> checked the IntelliJ log and found some error message like: >> >> 2019-04-14 16:05:24,796 [ 314609] INFO - #org.jetbrains.idea.maven - >> [WARNING] Rule 0: org.apache.maven.plugins.enforcer.RequireMavenVersion >> failed with message: >> Detected Maven Version: 3.3.9 is not in the allowed range 3.6.0. >> 2019-04-14 16:05:24,813 [ 314626] INFO - #org.jetbrains.idea.maven - >> org.apache.maven.lifecycle.LifecycleExecutionException: Failed to execute >> goal org.apache.maven.plugins:maven-enforcer-plugin:3.0.0-M2:enforce >> (enforce-versions) on project spark-parent_2.12: Some Enforcer rules have >> failed. Look above for specific messages explaining why the rule failed. >> java.lang.RuntimeException: >> org.apache.maven.lifecycle.LifecycleExecutionException: Failed to execute >> goal org.apache.maven.plugins:maven-enforcer-plugin:3.0.0-M2:enforce >> (enforce-versions) on project spark-parent_2.12: Some Enforcer rules have >> failed. Look above for specific messages explaining why the rule failed. >> >> >> It seems that the source file generation process failed sliently due to >> incorrect maven version. My IntelliJ bundled maven 3.3.9 while Spark's >> master branch requires maven 3.6.0. Be honest, failing an action sliently >> should be an IntelliJ bug. But a note to guide spark developers to set the >> maven version for IntelliJ should be helpful. I just create a JIRA >> (https://issues.apache.org/jira/browse/SPARK-27458) for this. >> >> Thanks and regards, >> William >> >> On Sun, 14 Apr 2019 at 8:57 PM, Sean Owen <sro...@gmail.com> wrote: >>> >>> For IntelliJ, in the Maven pane, there's a button to generate all >>> sources and resources that the build creates. That's the easier >>> option. You can open a PR to add a note about it along with other docs >>> for IntelliJ users. >>> >>> On Sun, Apr 14, 2019 at 4:24 AM William Wong <william1...@gmail.com> wrote: >>> > >>> > Dear all, >>> > >>> > I tried to follow the guide at >>> > 'http://spark.apache.org/developer-tools.html' to setup an IntelliJ >>> > project for Spark. However, the project was failed to build. It was due >>> > to missing classes generated via antlr on sql/catalyst project. >>> > >>> > I would like to enhance the document to hint other new joiners to run >>> > 'build/mvn antlr4 -f sql/catalyst/pom.xml' if hitting missing ANTLR4 >>> > classes files. However, Spark's project structure is very new to me. I >>> > hope I did not miss any guideline on spark's documentation about this >>> > issue, if so pls let me know. Thanks in advance. >>> > >>> > Regards, >>> > William >>> > >>> > --------------------------------------------------------------------- To unsubscribe e-mail: dev-unsubscr...@spark.apache.org