As Shane wrote, not yet.

`one build for works for both` is our aspiration and the next step
mentioned in the first email.

> The next step is `how to support JDK8/JDK11 together in a single
artifact`.

For the downstream users who build from the Apache Spark source, that will
not be a blocker because they will prefer a single JDK.

Bests,
Dongjoon.

On Mon, Aug 26, 2019 at 10:28 AM Shane Knapp <skn...@berkeley.edu> wrote:

> maybe in the future, but not right now as the hadoop 2.7 build is broken.
>
> also, i busted dev/run-tests.py in my changes to support java11 in PRBs:
> https://github.com/apache/spark/pull/25585
>
> quick fix, testing now.
>
> On Mon, Aug 26, 2019 at 10:23 AM Reynold Xin <r...@databricks.com> wrote:
>
>> Would it be possible to have one build that works for both?
>>
>>
>

Reply via email to