No guru to enlighten/highlighten ;) , but here's what I think - This is a balance call between the in-memory and disk usage.If you think the mappers are high which can again be a function of your input size, block size, you can tune further parameters during map/reduce steps to avoid expensive IO operations.
Have you looked into the parameters on the hadoop tutorial page? http://hadoop.apache.org/common/docs/current/mapred_tutorial.html#Map+Parameters http://hadoop.apache.org/common/docs/current/mapred_tutorial.html#Shuffle%2FReduce+Parameters Cheers, / On 4/15/10 4:19 PM, "eltonsky" <eltont9...@hotmail.com> wrote: Hello everyone, I know when map func generates intermediate output, reduce func will pull data directly from all maps' local disk. Although we can use combiner func to minimize the amount of data, when we have many mappers, say 10,000, that will be a crazy IO headache. And that dosen't seem right. Can anyone highlighten me on this? Regards, Elton -- View this message in context: http://old.nabble.com/What-it%27s-gonna-happend-when-it-comes-to-large-number-of-maps--tp28253798p28253798.html Sent from the Hadoop core-dev mailing list archive at Nabble.com.