As Shane wrote, not yet. `one build for works for both` is our aspiration and the next step mentioned in the first email.
> The next step is `how to support JDK8/JDK11 together in a single artifact`. For the downstream users who build from the Apache Spark source, that will not be a blocker because they will prefer a single JDK. Bests, Dongjoon. On Mon, Aug 26, 2019 at 10:28 AM Shane Knapp <skn...@berkeley.edu> wrote: > maybe in the future, but not right now as the hadoop 2.7 build is broken. > > also, i busted dev/run-tests.py in my changes to support java11 in PRBs: > https://github.com/apache/spark/pull/25585 > > quick fix, testing now. > > On Mon, Aug 26, 2019 at 10:23 AM Reynold Xin <r...@databricks.com> wrote: > >> Would it be possible to have one build that works for both? >> >> >