i've realized that for each link, i spend most of the time in the following
perl script
foreach my $url (@lines){ -- I READ MY 1-ROW URL FILE
$contador=2;
$test=0;
while(!$test){
$browser2->get($url);
$content = $browser2->content();
--IN THESE 2 STEPS I SPEND 6 SECONDS for a 86 kb html, Is it ok? Can i
perform these 2 steps faster?
Thanks!
T
On 1/22/07, Igor Sutton <[EMAIL PROTECTED]> wrote:
Hi Tatiana,
2007/1/22, Tatiana Lloret Iglesias <[EMAIL PROTECTED]>:
> I do it from the JAva and not from the PErl because i need to perform an
> insert into the database each time i process a link and also i have to
> inform via rss about the progress of the global download process (23.343out
> of 70.000 files have been downloaded) ....
>
You can use the awesome XML::RSS module to create RSS. Now, for
database insert you have the excellent DBI module.
I bet Java is your problem there (no, I'm not initiating a language war
here).
--
Igor Sutton Lopes <[EMAIL PROTECTED]>