Hey Andrew,

Indeed, sometimes there are patches that sit around a while and in
this case it can be because it's unclear to the reviewers whether they
are features worth having - or just by accident.

To put things in perspective, Spark merges about 80% of the proposed
patches (if you look we are on around 600 since moving to the new repo
with 100 not merged) - so in general we try hard to be very supportive
of community patches, much more than other projects in this space.

- Patrick

On Mon, Feb 24, 2014 at 1:39 PM, Matei Zaharia <matei.zaha...@gmail.com> wrote:
> Thanks for bringing this up. One issue that makes this harder is that old 
> inactive PRs on GitHub are not really getting closed, so active ones might be 
> lost between those. For now please just post on the dev list if your PR is 
> being ignored. We'll implement some kind of cleanup (at least manually) to 
> close the old ones.
>
> Matei
>
> On Feb 24, 2014, at 1:30 PM, Andrew Ash <and...@andrewash.com> wrote:
>
>> Yep that's the one thanks! That's quite a few more people than I thought
>>
>> Sent from my mobile phone
>> On Feb 24, 2014 1:20 PM, "Nan Zhu" <zhunanmcg...@gmail.com> wrote:
>>
>>> Do you mean this
>>> https://cwiki.apache.org/confluence/display/SPARK/Committers?
>>>
>>> --
>>> Nan Zhu
>>>
>>>
>>> On Monday, February 24, 2014 at 4:18 PM, Andrew Ash wrote:
>>>
>>>> Would love to have a discussion since I know the core contributors are
>>>> facing a barrage of PRs and things are falling through the cracks.
>>>>
>>>> Is there a list of who can commit to core Spark somewhere? Maybe that
>>> list
>>>> should be expanded or there should be a rotation of PR duty of some sort.
>>>>
>>>> One of the perils of having a vibrant, organic community is that you get
>>>> way more contributions than you expected!
>>>>
>>>>
>>>> On Mon, Feb 24, 2014 at 1:16 PM, Nan Zhu <zhunanmcg...@gmail.com(mailto:
>>> zhunanmcg...@gmail.com)> wrote:
>>>>
>>>>> yet another email about forgotten PR
>>>>>
>>>>> I think Sean would like to start some discussion on the current
>>> situation
>>>>> where committers are facing a flood of PRs recently (as he said in the
>>>>> discussion thread about how to prevent the blob of RDD API)?
>>>>>
>>>>> Best,
>>>>>
>>>>> --
>>>>> Nan Zhu
>>>>>
>>>>>
>>>>> On Monday, February 24, 2014 at 4:07 PM, Andrew Ash wrote:
>>>>>
>>>>>> Hi Spark devs,
>>>>>>
>>>>>> Kyle identified a deficiency in Spark where generating iterators are
>>>>>> unrolled into memory and then flushed to disk rather than sent
>>> straight
>>>>>>
>>>>>
>>>>> to
>>>>>> disk when possible.
>>>>>>
>>>>>> He's had a patch sitting ready for code review for quite some time
>>> now
>>>>> (100
>>>>>> days) but no response.
>>>>>>
>>>>>> Is this something that an admin would be able to review? I for one
>>> would
>>>>>> find this quite valuable.
>>>>>>
>>>>>> Thanks!
>>>>>> Andrew
>>>>>>
>>>>>>
>>>>>> https://spark-project.atlassian.net/browse/SPARK-942
>>>>>> https://github.com/apache/incubator-spark/pull/180
>>>>>>
>>>>>
>>>>>
>>>>
>>>>
>>>>
>>>
>>>
>>>
>

Reply via email to