Hi, These happens when operators are used in queries (Hive Operators). Hive creates 4 counters per operator, max upto 1000, plus a few additional counters like file read/write, partitions and tables. Hence the number of counter required is going to be dependent upon the query.
Using "EXPLAIN EXTENDED" and "grep -ri operators | wc -l" print out the used numbers of operators. Use this value to tweak the MR settings carefully. Praveen has a good explanation 'bout counters online: http://www.thecloudavenue.com/2011/12/limiting-usage-counters-in-hadoop.html Rule of thumb for Hive: count of operators * 4 + n (n for file ops and other stuff). cheers, Alex On Jan 2, 2013, at 10:35 AM, Krishna Rao <krishnanj...@gmail.com> wrote: > A particular query that I run fails with the following error: > > *** > Job 18: Map: 2 Reduce: 1 Cumulative CPU: 3.67 sec HDFS Read: 0 HDFS > Write: 0 SUCCESS > Exception in thread "main" > org.apache.hadoop.mapreduce.counters.LimitExceededException: Too many > counters: 121 max=120 > ... > *** > > Googling suggests that I should increase "mapreduce.job.counters.limit". > And that the number of counters a job uses > has an effect on the memory used by the JobTracker, so I shouldn't increase > this number too high. > > Is there a rule of thumb for what this number should be as a function of > JobTracker memory? That is should I be cautious and > increase by 5 at a time, or could I just double it? > > Cheers, > > Krishna -- Alexander Alten-Lorenz http://mapredit.blogspot.com German Hadoop LinkedIn Group: http://goo.gl/N8pCF