Hi, is it possible to avoid some memory overhead with a mongodb cursor and multiprocessing? Regarding to the size of the cursor, Python consuming at first a lot of memory. However the estimation is independend among each other document (chunking?).
Maybe there is a better way using multiprocessing in place of Pool? score_proc_pool.map(scoring_wrapper,mongo_cursor,chunksize=10000) Inside the scoring_wrapper I'm writing estimated scores without a return value. def scoring_wrapper(doc): ........ profiles.update({'anyid':anyid}, { '$set':{'profile':value} },upsert=True) Thanks in advance Christian -- http://mail.python.org/mailman/listinfo/python-list