My fear with logging the complete data input is that it will make the
problem worse for my customers because this problem is only happening on
heavily loaded servers. I can't reproduce it locally.

On Tue, Sep 6, 2016 at 11:26 AM, Perrin Harkins <phark...@gmail.com> wrote:

> Hi John,
>
> The key is usually finding out what the request was that caused it. You
> can add the pid to your access logging, or write a more complete mod_perl
> handler to log the complete data input along with the pid. Then you just go
> back and look at what it was after you see which process was killed.
>
> - Perrin
>
> On Tue, Sep 6, 2016 at 10:00 AM, John Dunlap <j...@lariat.co> wrote:
>
>> The system load reported by the uptime command, on one of my servers,
>> periodically spikes to 20-30 and then, shortly thereafter, I see this in
>> dmesg:
>>
>> [2887460.393402] Out of memory: Kill process 12533 (/usr/sbin/apach)
>> score 25 or sacrifice child
>> [2887460.394880] Killed process 12533 (/usr/sbin/apach)
>> total-vm:476432kB, anon-rss:204480kB, file-rss:0kB
>>
>> Several gigs of memory then becomes available and the system load quickly
>> returns to normal. I'm pretty sure it's a mod perl process that's doing
>> this but I'm not entirely sure how to track down the problem.
>>
>> How would you guys approach this problem?
>>
>> --
>> John Dunlap
>> *CTO | Lariat *
>>
>> *Direct:*
>> *j...@lariat.co <j...@lariat.co>*
>>
>> *Customer Service:*
>> 877.268.6667
>> supp...@lariat.co
>>
>
>


-- 
John Dunlap
*CTO | Lariat *

*Direct:*
*j...@lariat.co <j...@lariat.co>*

*Customer Service:*
877.268.6667
supp...@lariat.co

Reply via email to