Would it be possible to have one build that works for both?

On Mon, Aug 26, 2019 at 10:22 AM Dongjoon Hyun <dongjoon.h...@gmail.com>
wrote:

> Thank you all!
>
> Let me add more explanation on the current status.
>
>     - If you want to run on JDK8, you need to build on JDK8
>     - If you want to run on JDK11, you need to build on JDK11.
>
> The other combinations will not work.
>
> Currently, we have two Jenkins jobs. (1) is the one I pointed, and (2) is
> the one for the remaining community work.
>
>     1) Build and test on JDK11 (spark-master-test-maven-hadoop-3.2-jdk-11)
>     2) Build on JDK8 and test on JDK11
> (spark-master-test-maven-hadoop-2.7-jdk-11-ubuntu-testing)
>
> To keep JDK11 compatibility, the following is merged today.
>
>     [SPARK-28701][TEST-HADOOP3.2][TEST-JAVA11][K8S] adding java11
> support for pull request builds
>
> But, we still have many stuffs to do for Jenkins/Release and we need your
> support about JDK11. :)
>
> Bests,
> Dongjoon.
>
>
> On Sun, Aug 25, 2019 at 10:30 PM Takeshi Yamamuro <linguin....@gmail.com>
> wrote:
>
>> Cool, congrats!
>>
>> Bests,
>> Takeshi
>>
>> On Mon, Aug 26, 2019 at 1:01 PM Hichame El Khalfi <hich...@elkhalfi.com>
>> wrote:
>>
>>> That's Awesome !!!
>>>
>>> Thanks to everyone that made this possible :cheers:
>>>
>>> Hichame
>>>
>>> *From:* cloud0...@gmail.com
>>> *Sent:* August 25, 2019 10:43 PM
>>> *To:* lix...@databricks.com
>>> *Cc:* felixcheun...@hotmail.com; ravishankar.n...@gmail.com;
>>> dongjoon.h...@gmail.com; dev@spark.apache.org; u...@spark.apache.org
>>> *Subject:* Re: JDK11 Support in Apache Spark
>>>
>>> Great work!
>>>
>>> On Sun, Aug 25, 2019 at 6:03 AM Xiao Li <lix...@databricks.com> wrote:
>>>
>>>> Thank you for your contributions! This is a great feature for Spark
>>>> 3.0! We finally achieve it!
>>>>
>>>> Xiao
>>>>
>>>> On Sat, Aug 24, 2019 at 12:18 PM Felix Cheung <
>>>> felixcheun...@hotmail.com> wrote:
>>>>
>>>>> That’s great!
>>>>>
>>>>> ------------------------------
>>>>> *From:* ☼ R Nair <ravishankar.n...@gmail.com>
>>>>> *Sent:* Saturday, August 24, 2019 10:57:31 AM
>>>>> *To:* Dongjoon Hyun <dongjoon.h...@gmail.com>
>>>>> *Cc:* dev@spark.apache.org <dev@spark.apache.org>; user @spark/'user
>>>>> @spark'/spark users/user@spark <u...@spark.apache.org>
>>>>> *Subject:* Re: JDK11 Support in Apache Spark
>>>>>
>>>>> Finally!!! Congrats
>>>>>
>>>>> On Sat, Aug 24, 2019, 11:11 AM Dongjoon Hyun <dongjoon.h...@gmail.com>
>>>>> wrote:
>>>>>
>>>>>> Hi, All.
>>>>>>
>>>>>> Thanks to your many many contributions,
>>>>>> Apache Spark master branch starts to pass on JDK11 as of today.
>>>>>> (with `hadoop-3.2` profile: Apache Hadoop 3.2 and Hive 2.3.6)
>>>>>>
>>>>>>
>>>>>> https://amplab.cs.berkeley.edu/jenkins/view/Spark%20QA%20Test%20(Dashboard)/job/spark-master-test-maven-hadoop-3.2-jdk-11/326/
>>>>>>     (JDK11 is used for building and testing.)
>>>>>>
>>>>>> We already verified all UTs (including PySpark/SparkR) before.
>>>>>>
>>>>>> Please feel free to use JDK11 in order to build/test/run `master`
>>>>>> branch and
>>>>>> share your experience including any issues. It will help Apache Spark
>>>>>> 3.0.0 release.
>>>>>>
>>>>>> For the follow-ups, please follow
>>>>>> https://issues.apache.org/jira/browse/SPARK-24417 .
>>>>>> The next step is `how to support JDK8/JDK11 together in a single
>>>>>> artifact`.
>>>>>>
>>>>>> Bests,
>>>>>> Dongjoon.
>>>>>>
>>>>>
>>>>
>>>> --
>>>> [image: Databricks Summit - Watch the talks]
>>>> <https://databricks.com/sparkaisummit/north-america>
>>>>
>>>
>>
>> --
>> ---
>> Takeshi Yamamuro
>>
>

Reply via email to