If we go for RC2, we should include both: https://github.com/apache/spark/pull/27210 https://github.com/apache/spark/pull/27184
just for the sake of being complete and making the maintenance simple. 2020년 1월 16일 (목) 오후 12:38, Wenchen Fan <cloud0...@gmail.com>님이 작성: > Recently we merged several fixes to 2.4: > https://issues.apache.org/jira/browse/SPARK-30325 a driver hang issue > https://issues.apache.org/jira/browse/SPARK-30246 a memory leak issue > https://issues.apache.org/jira/browse/SPARK-29708 a correctness > issue(for a rarely used feature, so not merged to 2.4 yet) > > Shall we include them? > > > On Wed, Jan 15, 2020 at 9:51 PM Hyukjin Kwon <gurwls...@gmail.com> wrote: > >> +1 >> >> On Wed, 15 Jan 2020, 08:24 Takeshi Yamamuro, <linguin....@gmail.com> >> wrote: >> >>> +1; >>> >>> I checked the links and materials, then I run the tests with >>> `-Pyarn -Phadoop-2.7 -Phive -Phive-thriftserver -Pmesos -Pkubernetes >>> -Psparkr` >>> on macOS (Java 8). >>> All the things look fine and I didn't see the error on my env >>> that Sean said above. >>> >>> Thanks, Dongjoon! >>> >>> Bests, >>> Takeshi >>> >>> On Wed, Jan 15, 2020 at 4:09 AM DB Tsai <dbt...@dbtsai.com> wrote: >>> >>>> +1 Thanks. >>>> >>>> Sincerely, >>>> >>>> DB Tsai >>>> ---------------------------------------------------------- >>>> Web: https://www.dbtsai.com >>>> PGP Key ID: 42E5B25A8F7A82C1 >>>> >>>> On Tue, Jan 14, 2020 at 11:08 AM Sean Owen <sro...@apache.org> wrote: >>>> > >>>> > Yeah it's something about the env I spun up, but I don't know what. It >>>> > happens frequently when I test, but not on Jenkins. >>>> > The Kafka error comes up every now and then and a clean rebuild fixes >>>> > it, but not in my case. I don't know why. >>>> > But if nobody else sees it, I'm pretty sure it's just an artifact of >>>> > the local VM. >>>> > >>>> > On Tue, Jan 14, 2020 at 12:57 PM Dongjoon Hyun < >>>> dongjoon.h...@gmail.com> wrote: >>>> > > >>>> > > Thank you, Sean. >>>> > > >>>> > > First of all, the `Ubuntu` job on Amplab Jenkins farm is green. >>>> > > >>>> > > >>>> https://amplab.cs.berkeley.edu/jenkins/view/Spark%20QA%20Test/job/spark-branch-2.4-test-sbt-hadoop-2.7-ubuntu-testing/ >>>> > > >>>> > > For the failures, >>>> > > 1. Yes, the `HiveExternalCatalogVersionsSuite` flakiness is a >>>> known one. >>>> > > 2. For `HDFSMetadataLogSuite` failure, I also observed a few >>>> time before in CentOS too. >>>> > > 3. Kafka build error is new to me. Does it happen on `Maven` >>>> clean build? >>>> > > >>>> > > Bests, >>>> > > Dongjoon. >>>> > > >>>> > > >>>> > > On Tue, Jan 14, 2020 at 6:40 AM Sean Owen <sro...@apache.org> >>>> wrote: >>>> > >> >>>> > >> +1 from me. I checked sigs/licenses, and built/tested from source >>>> on >>>> > >> Java 8 + Ubuntu 18.04 with " -Pyarn -Phive -Phive-thriftserver >>>> > >> -Phadoop-2.7 -Pmesos -Pkubernetes -Psparkr -Pkinesis-asl". I do get >>>> > >> test failures, but, these are some I have always seen on Ubuntu, >>>> and I >>>> > >> do not know why they happen. They don't seem to affect others, but, >>>> > >> let me know if anyone else sees these? >>>> > >> >>>> > >> >>>> > >> Always happens for me: >>>> > >> >>>> > >> - HDFSMetadataLog: metadata directory collision *** FAILED *** >>>> > >> The await method on Waiter timed out. >>>> (HDFSMetadataLogSuite.scala:178) >>>> > >> >>>> > >> This one has been flaky at times due to external dependencies: >>>> > >> >>>> > >> org.apache.spark.sql.hive.HiveExternalCatalogVersionsSuite *** >>>> ABORTED *** >>>> > >> Exception encountered when invoking run on a nested suite - >>>> > >> spark-submit returned with exit code 1. >>>> > >> Command line: './bin/spark-submit' '--name' 'prepare testing >>>> tables' >>>> > >> '--master' 'local[2]' '--conf' 'spark.ui.enabled=false' '--conf' >>>> > >> 'spark.master.rest.enabled=false' '--conf' >>>> > >> >>>> 'spark.sql.warehouse.dir=/data/spark-2.4.5/sql/hive/target/tmp/warehouse-c2f762fd-688e-42b7-a822-06823a6bbd98' >>>> > >> '--conf' 'spark.sql.test.version.index=0' '--driver-java-options' >>>> > >> >>>> '-Dderby.system.home=/data/spark-2.4.5/sql/hive/target/tmp/warehouse-c2f762fd-688e-42b7-a822-06823a6bbd98' >>>> > >> '/data/spark-2.4.5/sql/hive/target/tmp/test7297526474581770293.py' >>>> > >> >>>> > >> Kafka doesn't build with this weird error. I tried a clean build. I >>>> > >> think we've seen this before. >>>> > >> >>>> > >> [error] This symbol is required by 'method >>>> > >> org.apache.spark.metrics.MetricsSystem.getServletHandlers'. >>>> > >> [error] Make sure that term eclipse is in your classpath and check >>>> for >>>> > >> conflicting dependencies with `-Ylog-classpath`. >>>> > >> [error] A full rebuild may help if 'MetricsSystem.class' was >>>> compiled >>>> > >> against an incompatible version of org. >>>> > >> [error] testUtils.sendMessages(topic, data.toArray) >>>> > >> [error] >>>> > >> >>>> > >> On Mon, Jan 13, 2020 at 6:28 AM Dongjoon Hyun < >>>> dongjoon.h...@gmail.com> wrote: >>>> > >> > >>>> > >> > Please vote on releasing the following candidate as Apache Spark >>>> version 2.4.5. >>>> > >> > >>>> > >> > The vote is open until January 16th 5AM PST and passes if a >>>> majority +1 PMC votes are cast, with a minimum of 3 +1 votes. >>>> > >> > >>>> > >> > [ ] +1 Release this package as Apache Spark 2.4.5 >>>> > >> > [ ] -1 Do not release this package because ... >>>> > >> > >>>> > >> > To learn more about Apache Spark, please see >>>> http://spark.apache.org/ >>>> > >> > >>>> > >> > The tag to be voted on is v2.4.5-rc1 (commit >>>> 33bd2beee5e3772a9af1d782f195e6a678c54cf0): >>>> > >> > https://github.com/apache/spark/tree/v2.4.5-rc1 >>>> > >> > >>>> > >> > The release files, including signatures, digests, etc. can be >>>> found at: >>>> > >> > https://dist.apache.org/repos/dist/dev/spark/v2.4.5-rc1-bin/ >>>> > >> > >>>> > >> > Signatures used for Spark RCs can be found in this file: >>>> > >> > https://dist.apache.org/repos/dist/dev/spark/KEYS >>>> > >> > >>>> > >> > The staging repository for this release can be found at: >>>> > >> > >>>> https://repository.apache.org/content/repositories/orgapachespark-1339/ >>>> > >> > >>>> > >> > The documentation corresponding to this release can be found at: >>>> > >> > https://dist.apache.org/repos/dist/dev/spark/v2.4.5-rc1-docs/ >>>> > >> > >>>> > >> > The list of bug fixes going into 2.4.5 can be found at the >>>> following URL: >>>> > >> > https://issues.apache.org/jira/projects/SPARK/versions/12346042 >>>> > >> > >>>> > >> > This release is using the release script of the tag v2.4.5-rc1. >>>> > >> > >>>> > >> > FAQ >>>> > >> > >>>> > >> > ========================= >>>> > >> > How can I help test this release? >>>> > >> > ========================= >>>> > >> > >>>> > >> > If you are a Spark user, you can help us test this release by >>>> taking >>>> > >> > an existing Spark workload and running on this release >>>> candidate, then >>>> > >> > reporting any regressions. >>>> > >> > >>>> > >> > If you're working in PySpark you can set up a virtual env and >>>> install >>>> > >> > the current RC and see if anything important breaks, in the >>>> Java/Scala >>>> > >> > you can add the staging repository to your projects resolvers >>>> and test >>>> > >> > with the RC (make sure to clean up the artifact cache >>>> before/after so >>>> > >> > you don't end up building with a out of date RC going forward). >>>> > >> > >>>> > >> > =========================================== >>>> > >> > What should happen to JIRA tickets still targeting 2.4.5? >>>> > >> > =========================================== >>>> > >> > >>>> > >> > The current list of open tickets targeted at 2.4.5 can be found >>>> at: >>>> > >> > https://issues.apache.org/jira/projects/SPARK and search for >>>> "Target Version/s" = 2.4.5 >>>> > >> > >>>> > >> > Committers should look at those and triage. Extremely important >>>> bug >>>> > >> > fixes, documentation, and API tweaks that impact compatibility >>>> should >>>> > >> > be worked on immediately. Everything else please retarget to an >>>> > >> > appropriate release. >>>> > >> > >>>> > >> > ================== >>>> > >> > But my bug isn't fixed? >>>> > >> > ================== >>>> > >> > >>>> > >> > In order to make timely releases, we will typically not hold the >>>> > >> > release unless the bug in question is a regression from the >>>> previous >>>> > >> > release. That being said, if there is something which is a >>>> regression >>>> > >> > that has not been correctly targeted please ping me or a >>>> committer to >>>> > >> > help target the issue. >>>> > >>>> > --------------------------------------------------------------------- >>>> > To unsubscribe e-mail: dev-unsubscr...@spark.apache.org >>>> > >>>> >>>> --------------------------------------------------------------------- >>>> To unsubscribe e-mail: dev-unsubscr...@spark.apache.org >>>> >>>> >>> >>> -- >>> --- >>> Takeshi Yamamuro >>> >>