Wells Oliver wrote:
Writing a class which essentially spiders a site and saves the files
locally. On a URLError exception, it sleeps for a second and tries again (on
404 it just moves on). The relevant bit of code, including the offending
method:
class Handler(threading.Thread):
def __init__(self, url):
threading.Thread.__init__(self)
self.url = url
def save(self, uri, location):
try:
handler = urllib2.urlopen(uri)
except urllib2.HTTPError, e:
if e.code == 404:
return
else:
print "retrying %s (HTTPError)" % uri
time.sleep(1)
self.save(uri, location)
except urllib2.URLError, e:
print "retrying %s" % uri
time.sleep(1)
self.save(uri, location)
if not os.path.exists(os.path.dirname(location)):
os.makedirs(os.path.dirname(location))
file = open(location, "w")
file.write(handler.read())
file.close()
...
But what I am seeing is that after a retry (on catching a URLError
exception), I see bunches of "UnboundLocalError: local variable 'handler'
referenced before assignment" errors on line 38, which is the
"file.write(handler.read())" line..
What gives?
Your save() method is recursive in the case of that error, which is a
poor excuse for what should have been a loop. You should have some
retry depth check, just in case you get hundreds of such errors on the
same site.
But ignoring that issue, the specific symptom is caused when returning
from the recursive call. You still fall through to the end of your
outer method call, and try to write stuff from that handler (also wiping
out the file "location" in the process). Without a handler, that causes
an exception. So you should follow those calls to self.save() with
return's.
Much better would be to write a loop in the first place, and break out
of the loop when you succeed. Then the flow is much easier to follow,
and you can easily avoid the risk of looping forever, nor of running out
of stack space.
Something like (untested):
def save(self, uri, location):
for tries in xrange(10): #try up to 10 times
try:
handler = urllib2.urlopen(uri)
except urllib2.HTTPError, e:
if e.code == 404:
break
else:
print "retrying %s (HTTPError)" % uri
time.sleep(1)
continue
except urllib2.URLError, e:
print "retrying %s" % uri
time.sleep(1)
continue
if not os.path.exists(os.path.dirname(location)):
os.makedirs(os.path.dirname(location))
file = open(location, "w")
file.write(handler.read())
file.close()
break
Other refinements are possible, of course. For example, if you have any
cleanup code at the end you may need an additional flag variable to tell
whether you've succeeded or not.
--
http://mail.python.org/mailman/listinfo/python-list