Thanks Peter - good point > -----Original Message----- > From: Peter Scott [mailto:[EMAIL PROTECTED] > Sent: Sunday, February 23, 2003 5:17 AM > To: [EMAIL PROTECTED] > Subject: RE: Out of memory while finding duplicate rows > > > In article <[EMAIL PROTECTED]>, > [EMAIL PROTECTED] (Beau E. Cox) writes: > >Hi - > > > >Wait! If you are going to load the data into a database anyway, > >why not use the existing database (or the one being created) to > >remove duplicates. You don't even have to have an index on the > >column you are making unique (but it would be _much_ faster). > >Just select on you key, and, if found, reject the datum as > >a duplicate. You really shouldn't have to go to any draconian > >measures to find duplicates! > > No need even to do that. Just set a primary key constraint on > the database table (like all good tables should have anyway) > and you're done. (Or if the "duplicate" criterion involves > some other column, put a UNIQUE constraint on it.) Then all > inserts of duplicate records will fail automatically. Just > make sure that RaiseError is set to false in the DBI connect. > > -- > Peter Scott > http://www.perldebugged.com
I was speaking in general terms - I have no idea what the structure of his target db is, so my manual way covers all bases... :) Aloha => Beau; -- To unsubscribe, e-mail: [EMAIL PROTECTED] For additional commands, e-mail: [EMAIL PROTECTED]