On Mar 14, 3:26 pm, [EMAIL PROTECTED] (Manoj) wrote:
> When using Data: Dumper is taking more time for my 10000 lines of CSV file.
> This solved a few queries...and the benchmark was a new value addition for
> me. Thanks
> > 2)       Is there any optimal method for reading a CSV file and put to
> hash table.
>
> > May have better approaches however this read CSV into Hash
>
> > use Data::Dumper;
> > open(INFILE, "<", "sample.csv") or die $!;
> > my %hsh;
> > %hsh = ( %hsh, (split(/,/, $_))[1,2] ) while ( <INFILE> );
>
> That is a *very* inefficient way to populate a hash as you are copying
> the entire hash for every record in the file.  Better to add the keys
> and values individually:
>
> my %hsh;
> while ( <INFILE> ) {
>      chomp;
>      my ( $key, $value ) = split /,/;
>      }
>
> > print Dumper \%hsh;
>
> John

> - Show quoted text -

Suppose my csv file has 5 columns: f id fa mo ge
Will my ($key, $value) = split/,/; still work?
Is there any other mothod that is more efficient?

JB


--
To unsubscribe, e-mail: [EMAIL PROTECTED]
For additional commands, e-mail: [EMAIL PROTECTED]
http://learn.perl.org/


Reply via email to