The reduce percentage keeps fluctuating when, the alter index command is being keyed. The logs just give " out of memory error " after tweaking some properties, the earlier exceptions doesn't appear now. Can anyone guide me here? I have increased the heap space upto 4gb.. still, getting the same exception..
On 28 July 2011 16:37, Siddharth Ramanan <siddharth.rama...@gmail.com>wrote: > Hi, > I am adding the log information for a reduce task. I am running hadoop > in standalone mode. > > 2011-07-28 19:16:42,621 ERROR > org.apache.hadoop.hive.ql.stats.jdbc.JDBCStatsPublisher: Error during JDBC > connection to jdbc:derby:;databaseName=TempStatsStore;create=true. > java.lang.ClassNotFoundException: org.apache.derby.jdbc.EmbeddedDriver > at java.net.URLClassLoader$1.run(URLClassLoader.java:202) > at java.security.AccessController.doPrivileged(Native Method) > at java.net.URLClassLoader.findClass(URLClassLoader.java:190) > at java.lang.ClassLoader.loadClass(ClassLoader.java:306) > at sun.misc.Launcher$AppClassLoader.loadClass(Launcher.java:301) > at java.lang.ClassLoader.loadClass(ClassLoader.java:247) > at java.lang.Class.forName0(Native Method) > at java.lang.Class.forName(Class.java:169) > at > org.apache.hadoop.hive.ql.stats.jdbc.JDBCStatsPublisher.connect(JDBCStatsPublisher.java:55) > at > org.apache.hadoop.hive.ql.exec.FileSinkOperator.publishStats(FileSinkOperator.java:781) > at > org.apache.hadoop.hive.ql.exec.FileSinkOperator.closeOp(FileSinkOperator.java:649) > at org.apache.hadoop.hive.ql.exec.Operator.close(Operator.java:557) > at org.apache.hadoop.hive.ql.exec.Operator.close(Operator.java:566) > at org.apache.hadoop.hive.ql.exec.Operator.close(Operator.java:566) > at > org.apache.hadoop.hive.ql.exec.ExecReducer.close(ExecReducer.java:303) > at > org.apache.hadoop.mapred.ReduceTask.runOldReducer(ReduceTask.java:473) > at org.apache.hadoop.mapred.ReduceTask.run(ReduceTask.java:411) > at org.apache.hadoop.mapred.Child.main(Child.java:170) > 2011-07-28 19:16:42,622 ERROR > org.apache.hadoop.hive.ql.exec.FileSinkOperator: StatsPublishing error: > cannot connect to database > 2011-07-28 19:16:42,622 INFO > org.apache.hadoop.hive.ql.exec.FileSinkOperator: TABLE_ID_1_ROWCOUNT:126 > 2011-07-28 19:16:42,622 INFO org.apache.hadoop.hive.ql.exec.SelectOperator: > 5 Close done > 2011-07-28 19:16:42,622 INFO > org.apache.hadoop.hive.ql.exec.GroupByOperator: 4 Close done > 2011-07-28 19:16:42,625 INFO org.apache.hadoop.mapred.TaskRunner: > Task:attempt_201107271749_0029_r_000019_0 is done. And is in the process of > commiting > 2011-07-28 19:16:42,627 INFO org.apache.hadoop.mapred.TaskRunner: Task > 'attempt_201107271749_0029_r_000019_0' done. > > > On 28 July 2011 16:19, Siddharth Ramanan <siddharth.rama...@gmail.com>wrote: > >> Hi, >> I have a table, which has close to a billion rows.. I am trying to >> create an index for the table, when I do the alter command, I always end up >> with map-reduce jobs with errors. The same runs fine for small tables >> though, I also notice that the number of reducers are set to 24, even if set >> manually to 1 and also the reduce percentage changes in a bizarre way, it >> increases, then decreases and finally reaches 100 with a message job ended >> with errors. It would be useful, if I can get help on this.. >> >> Thanks, >> Siddharth >> > >