On Sunday, June 29, 2014 7:31:37 PM UTC+5:30, Roy Smith wrote: > In article <mailman.11325.1404048700.18130.python-l...@python.org>, > > Dave Angel <da...@davea.name> wrote: > > > > > subhabangal...@gmail.com Wrote in message: > > > > Dear Group, > > > > > > > > I am trying to crawl multiple URLs. As they are coming I want to write > > > them > > > > as string, as they are coming, preferably in a queue. > > > > > > > > If any one of the esteemed members of the group may kindly help. > > > > > > > > > > >From your subject line, it appears you want to keep multiple files open, > > > >and write to each in an arbitrary order. That's no problem, up to the > > > >operating system limits. Define a class that holds the URL information > > >and > > > >for each instance, add an attribute for an output file handle. > > > > > > Don't forget to close each file when you're done with the corresponding URL. > > > > One other thing to mention is that if you're doing anything with > > fetching URLs from Python, you almost certainly want to be using Kenneth > > Reitz's excellent requests module (http://docs.python-requests.org/). > > The built-in urllib support in Python works, but requests is so much > > simpler to use.
Dear Group, Sorry if I miscommunicated. I am opening multiple URLs with urllib.open, now one Url has huge html source files, like that each one has. As these files are read I am trying to concatenate them and put in one txt file as string. >From this big txt file I am trying to take out each html file body of each URL >and trying to write and store them with attempts like, for i, line in enumerate(file1): f = open("/python27/newfile_%i.txt" %i,'w') f.write(line) f.close() Generally not much of an issue, but was thinking of some better options. Regards, Subhabrata Banerjee. -- https://mail.python.org/mailman/listinfo/python-list