Thanks a lot. BR Loreia
On Thursday, August 16, 2012 1:23:39 PM UTC+2, Niphlod wrote: > > group by your unique columns, count the rows, find the ones with count > 1. > > db.define_table( > 'finddup', > Field('f1_name'), > Field('f2_name'), > Field('f3_name') > ) > fd = db.finddup > count = fd.id.count() > rtn = db(fd.id>0).select(fd.f1_name, fd.f2_name, fd.f3_name, count,groupby > =fd.f1_name|fd.f2_name|fd.f3_name, having=count>1) > > > > On Thursday, August 16, 2012 11:20:54 AM UTC+2, Loreia wrote: >> >> Hi, >> >> I am looking for a way to automatically find any possible duplicate >> records (rows) in one table of my Postgres database. >> Quick google-ing reveals it is easy to do it with pure SQL, e.g. one way >> specific for Postgres: >> http://stackoverflow.com/questions/1746213/how-to-delete-duplicate-entries-in-postgresql >> >> But I would like to be able to do it with web2p DAL (which would make it >> database independent). >> Is there a way to do it with web2py DAL? >> >> I just want to be able to find those duplicates, I don't need to delete >> them. >> >> Thank you and best regards, >> Loreia >> > --