Every hive query has a history file, and you can get these info from hive history file
Following java code can be an example: https://github.com/anjuke/hwi/blob/master/src/main/java/org/apache/hadoop/hive/hwi/util/QueryUtil.java Regard, Qiang 2013/1/30 Mathieu Despriee <mdespr...@octo.com> > Hi folks, > > I would like to run a list of generated HIVE queries. For each, I would > like to retrieve the MR job_id (or ids, in case of multiple stages). And > then, with this job_id, collect statistics from job tracker (cumulative > CPU, read bytes...) > > How can I send HIVE queries from a bash or python script, and retrieve the > job_id(s) ? > > For the 2nd part (collecting stats for the job), we're using a MRv1 Hadoop > cluster, so I don't have the AppMaster REST > API<http://hadoop.apache.org/docs/current/hadoop-yarn/hadoop-yarn-site/MapredAppMasterRest.html>. > I'm about to collect data from the jobtracker web UI. Any better idea ? > > Mathieu > > >