Thanks Fred for pointers...so far I was only able to build 2.1 with Java 7
and no zinc.
Will try options u suggest. FYI building with sbt ends up in oom even with
Java 7
I will try and update this thread
Kr

On 6 Oct 2016 8:58 pm, "Fred Reiss" <freiss....@gmail.com> wrote:

> There's no option to prevent build/mvn from starting the zinc server, but
> you should be able to prevent the maven build from using the zinc server by
> changing the <useZincServer> option at line 1935 of the master pom.xml.
>
> Note that the zinc-based compile works on my Ubuntu 16.04 box. You might
> be able to get zinc-based compiles working by tweaking your settings. A few
> things to try:
> -- Make sure another build hasn't left a second, incompatible copy of zinc
> squatting on the port that Spark expects to use
> -- Try setting the environment variable JAVA_7_HOME to point to and
> OpenJDK 7 installation. build/mvn runs zinc with Java 7 if that is
> available.
>
> Note that setting JAVA_7_HOME will break incremental compilation for
> sbt-based builds. Use that environment variable with restraint.
>
> Fred
>
>
>
> On Thu, Oct 6, 2016 at 2:22 AM, Marco Mistroni <mmistr...@gmail.com>
> wrote:
>
>> Thanks Fred
>> The build/mvn will trigger compilation using zinc and I want to avoid
>> that as every time I have tried it runs into errors while compiling spark
>> core. How can I disable zinc by default?
>> Kr
>>
>> On 5 Oct 2016 10:53 pm, "Fred Reiss" <freiss....@gmail.com> wrote:
>>
>>> Actually the memory options *are* required for Java 1.8. Without them
>>> the build will fail intermittently. We just updated the documentation with
>>> regard to this fact in Spark 2.0.1. Relevant PR is here:
>>> https://github.com/apache/spark/pull/15005
>>>
>>> Your best bet as the project transitions from Java 7 to Java 8 is to use
>>> the scripts build/mvn and build/sbt, which should be updated on a regular
>>> basis with safe JVM options.
>>>
>>> Fred
>>>
>>> On Wed, Oct 5, 2016 at 1:40 AM, Marco Mistroni <mmistr...@gmail.com>
>>> wrote:
>>>
>>>> Thanks Richard.  It also says that for Java 1.8 the mavenopts are not
>>>> required..unless I misinterpreted the instructions...
>>>> Kr
>>>>
>>>> On 5 Oct 2016 9:20 am, "Richard Siebeling" <rsiebel...@gmail.com>
>>>> wrote:
>>>>
>>>>> sorry, now with the link included, see http://spark.apache.org/do
>>>>> cs/latest/building-spark.html
>>>>>
>>>>> On Wed, Oct 5, 2016 at 10:19 AM, Richard Siebeling <
>>>>> rsiebel...@gmail.com> wrote:
>>>>>
>>>>>> Hi,
>>>>>>
>>>>>> did you set the following option: export MAVEN_OPTS="-Xmx2g
>>>>>> -XX:ReservedCodeCacheSize=512m"
>>>>>>
>>>>>> kind regards,
>>>>>> Richard
>>>>>>
>>>>>> On Tue, Oct 4, 2016 at 10:21 PM, Marco Mistroni <mmistr...@gmail.com>
>>>>>> wrote:
>>>>>>
>>>>>>> Hi all
>>>>>>>  my mvn build of Spark 2.1 using Java 1.8 is spinning out of memory
>>>>>>> with an error saying it cannot allocate enough memory during maven
>>>>>>> compilation
>>>>>>>
>>>>>>> Instructions (in the Spark 2.0 page) says that MAVENOPTS are not
>>>>>>> needed for Java 1.8 and , accoding to my understanding, spark build 
>>>>>>> process
>>>>>>> will add it
>>>>>>> during the build via mvn
>>>>>>> Note; i am not using Zinc. Rather, i am using my own Maven version
>>>>>>> (3.3.9), launching this command from the main spark directory. The same
>>>>>>> build works when i use Java 1.7(and MAVENOPTS)
>>>>>>>
>>>>>>> mvn -Pyarn -Dscala-2.11 -DskipTests clean package
>>>>>>>
>>>>>>> Could anyone assist?
>>>>>>> kr
>>>>>>>   marco
>>>>>>>
>>>>>>
>>>>>>
>>>>>
>>>
>

Reply via email to