An update: I cut the tag for RC1 last night. Currently fighting with the release process. Will post RC1 once I get it working.
On Tue, Mar 21, 2017 at 2:16 PM, Nick Pentreath <nick.pentre...@gmail.com> wrote: > As for SPARK-19759 <https://issues.apache.org/jira/browse/SPARK-19759>, I > don't think that needs to be targeted for 2.1.1 so we don't need to worry > about it > > > On Tue, 21 Mar 2017 at 13:49 Holden Karau <hol...@pigscanfly.ca> wrote: > >> I agree with Michael, I think we've got some outstanding issues but none >> of them seem like regression from 2.1 so we should be good to start the RC >> process. >> >> On Tue, Mar 21, 2017 at 1:41 PM, Michael Armbrust <mich...@databricks.com >> > wrote: >> >> Please speak up if I'm wrong, but none of these seem like critical >> regressions from 2.1. As such I'll start the RC process later today. >> >> On Mon, Mar 20, 2017 at 9:52 PM, Holden Karau <hol...@pigscanfly.ca> >> wrote: >> >> I'm not super sure it should be a blocker for 2.1.1 -- is it a >> regression? Maybe we can get TDs input on it? >> >> On Mon, Mar 20, 2017 at 8:48 PM Nan Zhu <zhunanmcg...@gmail.com> wrote: >> >> I think https://issues.apache.org/jira/browse/SPARK-19280 should be a >> blocker >> >> Best, >> >> Nan >> >> On Mon, Mar 20, 2017 at 8:18 PM, Felix Cheung <felixcheun...@hotmail.com> >> wrote: >> >> I've been scrubbing R and think we are tracking 2 issues >> >> https://issues.apache.org/jira/browse/SPARK-19237 >> >> https://issues.apache.org/jira/browse/SPARK-19925 >> >> >> >> >> ------------------------------ >> *From:* holden.ka...@gmail.com <holden.ka...@gmail.com> on behalf of >> Holden Karau <hol...@pigscanfly.ca> >> *Sent:* Monday, March 20, 2017 3:12:35 PM >> *To:* dev@spark.apache.org >> *Subject:* Outstanding Spark 2.1.1 issues >> >> Hi Spark Developers! >> >> As we start working on the Spark 2.1.1 release I've been looking at our >> outstanding issues still targeted for it. I've tried to break it down by >> component so that people in charge of each component can take a quick look >> and see if any of these things can/should be re-targeted to 2.2 or 2.1.2 & >> the overall list is pretty short (only 9 items - 5 if we only look at >> explicitly tagged) :) >> >> If your working on something for Spark 2.1.1 and it doesn't show up in >> this list please speak up now :) We have a lot of issues (including "in >> progress") that are listed as impacting 2.1.0, but they aren't targeted for >> 2.1.1 - if there is something you are working in their which should be >> targeted for 2.1.1 please let us know so it doesn't slip through the cracks. >> >> The query string I used for looking at the 2.1.1 open issues is: >> >> ((affectedVersion = 2.1.1 AND cf[12310320] is Empty) OR fixVersion = >> 2.1.1 OR cf[12310320] = "2.1.1") AND project = spark AND resolution = >> Unresolved ORDER BY priority DESC >> >> None of the open issues appear to be a regression from 2.1.0, but those >> seem more likely to show up during the RC process (thanks in advance to >> everyone testing their workloads :)) & generally none of them seem to be >> >> (Note: the cfs are for Target Version/s field) >> >> Critical Issues: >> SQL: >> SPARK-19690 <https://issues.apache.org/jira/browse/SPARK-19690> - Join >> a streaming DataFrame with a batch DataFrame may not work - PR >> https://github.com/apache/spark/pull/17052 (review in progress by >> zsxwing, currently failing Jenkins)* >> >> Major Issues: >> SQL: >> SPARK-19035 <https://issues.apache.org/jira/browse/SPARK-19035> - rand() >> function in case when cause failed - no outstanding PR (consensus on JIRA >> seems to be leaning towards it being a real issue but not necessarily >> everyone agrees just yet - maybe we should slip this?)* >> Deploy: >> SPARK-19522 <https://issues.apache.org/jira/browse/SPARK-19522> >> - --executor-memory flag doesn't work in local-cluster mode - >> https://github.com/apache/spark/pull/16975 (review in progress by >> vanzin, but PR currently stalled waiting on response) * >> Core: >> SPARK-20025 <https://issues.apache.org/jira/browse/SPARK-20025> - Driver >> fail over will not work, if SPARK_LOCAL* env is set. - >> https://github.com/apache/spark/pull/17357 (waiting on review) * >> PySpark: >> SPARK-19955 <https://issues.apache.org/jira/browse/SPARK-19955> - >> Update run-tests to support conda [ Part of Dropping 2.6 support -- which >> we shouldn't do in a minor release -- but also fixes pip installability >> tests to run in Jenkins ]- PR failing Jenkins (I need to poke this some >> more, but seems like 2.7 support works but some other issues. Maybe slip to >> 2.2?) >> >> Minor issues: >> Tests: >> SPARK-19612 <https://issues.apache.org/jira/browse/SPARK-19612> - Tests >> failing with timeout - No PR per-se but it seems unrelated to the 2.1.1 >> release. It's not targetted for 2.1.1 but listed as affecting 2.1.1 - I'd >> consider explicitly targeting this for 2.2? >> PySpark: >> SPARK-19570 <https://issues.apache.org/jira/browse/SPARK-19570> - Allow >> to disable hive in pyspark shell - https://github.com/apache/ >> spark/pull/16906 PR exists but its difficult to add automated tests for >> this (although if SPARK-19955 >> <https://issues.apache.org/jira/browse/SPARK-19955> gets in would make >> testing this easier) - no reviewers yet. Possible re-target?* >> Structured Streaming: >> SPARK-19613 <https://issues.apache.org/jira/browse/SPARK-19613> - Flaky >> test: StateStoreRDDSuite.versioning and immutability - It's not targetted >> for 2.1.1 but listed as affecting 2.1.1 - I'd consider explicitly targeting >> this for 2.2? >> ML: >> SPARK-19759 <https://issues.apache.org/jira/browse/SPARK-19759> >> - ALSModel.predict on Dataframes : potential optimization by not using >> blas - No PR consider re-targeting unless someone has a PR waiting in the >> wings? >> >> Explicitly targeted issues are marked with a *, the remaining issues are >> listed as impacting 2.1.1 and don't have a specific target version set. >> >> Since 2.1.1 continues the 2.1.0 branch, looking at 2.1.0 shows 1 open >> blocker in SQL( SPARK-19983 >> <https://issues.apache.org/jira/browse/SPARK-19983> ), >> >> Query string is: >> >> affectedVersion = 2.1.0 AND cf[12310320] is EMPTY AND project = spark AND >> resolution = Unresolved AND priority = targetPriority >> >> Continuing on for unresolved 2.1.0 issues in Major there are 163 (76 of >> them in progress), 65 Minor (26 in progress), and 9 trivial (6 in progress). >> >> I'll be going through the 2.1.0 major issues with open PRs that impact >> the PySpark component and seeing if any of them should be targeted for >> 2.1.1, if anyone from the other components wants to take a look through we >> might find some easy wins to be merged. >> >> Cheers, >> >> Holden :) >> >> -- >> Cell : 425-233-8271 <(425)%20233-8271> >> Twitter: https://twitter.com/holdenkarau >> >> >> -- >> Cell : 425-233-8271 <(425)%20233-8271> >> Twitter: https://twitter.com/holdenkarau >> >> >> >> >> >> -- >> Cell : 425-233-8271 <(425)%20233-8271> >> Twitter: https://twitter.com/holdenkarau >> >