Hi everyone,
I have a cluster where each container is configured at 4GB and some of my 
queries are getting over in 30 to 40 seconds. This leads me to believe that I 
have too much memory for my containers and I am thinking of reducing the 
container size to 1.5GB(hive.tez.container.size) but I am looking for a few 
more concrete data points to find out if really I have oversized containers?
I looked into the tez view of my DAG and the counters give me:
PHYSICAL_MEMORY_BYTES 907965628416
VIRTUAL_MEMORY_BYTES 1560263561216
I am guessing this is wrong as there is no way the query could finish in 20 
seconds on a 98GB cluster if the actual memory required by the query is 907GB. 
Any help to find some data points regarding determination of oversized 
containers is very much appreciated!

Thanks
Ranjan

Reply via email to