Also, a cold start for this task can take as little as 960 ms (where it
shows the log detail: "This request caused a new process to be started for
your application...").

And, the task is doing a URLFetch, so.. I'm sure that is taking up a chunk
of the running time.

One of my imports is pytz (no django or anything like that though).. so that
could also be of interest.

On Thu, Sep 23, 2010 at 12:10 PM, Eli Jones <[email protected]> wrote:

> My app is in Python.  The only thing it does between 6:30 AM PST and 1:00
> PM PST is run a single chained task.  The task just runs once every 5
> minutes.  (There are no other backend processes or users).
>
> Yesterday at 11:00 AM PST and 11:02 AM PST, the task experienced the 10
> second, 500 error (with 0 CPU_MS used).. then the task successfully retried
> at 11:02:55 AM PST.
>
> There was another one (not mentioned in your list of times for yesterday)
> at 7:45 AM PST.  Failed after 10 seconds with a 500 error and 0 CPU_MS used.
>  The task successfully retried at 7:47 AM PST (for some reason, the tasks in
> my app wait 120 seconds for the first retry)
>
> For some reason, the log doesn't contain the message that usually goes with
> this error (there is no detailed log message):
>
>  "Request was aborted after waiting too long to attempt to service your
> request. This may happen sporadically when the App Engine serving cluster is
> under unexpectedly high or uneven load. If you see this message frequently,
> please contact the App Engine team."
>
> But, it has the exact same fingerprint. (It just seems that the logs are
> trimming the detailed messages faster than usual.)
>
> This chained task runs 79 times each day (this would be 82 requests
> counting these 3 errors).. so these errors accounted for 3.66 % of the
> requests during the day yesterday.
>
>
> On Thu, Sep 23, 2010 at 11:19 AM, Ikai Lan (Google) <
> [email protected] <ikai.l%[email protected]>> wrote:
>
>> Hey guys,
>>
>> We've been tracking latency spikes and intermittent batches of aborted
>> requests over the last 24 hour period. There were at least 3 periods of
>> spikes yesterday:
>>
>> ~1100 Pacific (moderate, brief period of high)
>> ~1300 Pacific (small)
>> ~1530 Pacific (small)
>>
>> Community reports seem to indicate that these are related to loading
>> requests. If you are being affected, can you report your environment,
>> language and if these times correlate with your experiences in this thread?
>> Are there periods that this particular report did not cover? Please provide
>> measurements, times and application IDs.
>>
>> --
>> Ikai Lan
>> Developer Programs Engineer, Google App Engine
>> Blogger: http://googleappengine.blogspot.com
>> Reddit: http://www.reddit.com/r/appengine
>> Twitter: http://twitter.com/app_engine
>>
>>  --
>> You received this message because you are subscribed to the Google Groups
>> "Google App Engine" group.
>> To post to this group, send email to [email protected].
>> To unsubscribe from this group, send email to
>> [email protected]<google-appengine%[email protected]>
>> .
>> For more options, visit this group at
>> http://groups.google.com/group/google-appengine?hl=en.
>>
>
>

-- 
You received this message because you are subscribed to the Google Groups 
"Google App Engine" group.
To post to this group, send email to [email protected].
To unsubscribe from this group, send email to 
[email protected].
For more options, visit this group at 
http://groups.google.com/group/google-appengine?hl=en.

Reply via email to