thx, this worked now, after i checked, that i have to add them in each session. which config file have i to manipulate to embed those jars permantly?
2013/6/21 Ramki Palle <ramki.pa...@gmail.com> > Try to add the jar explicitly from hive prompt and see if that works. > > Regards, > Ramki. > > > On Fri, Jun 21, 2013 at 7:32 AM, fab wol <darkwoll...@gmail.com> wrote: > >> I'm using Hadoop 0.20.2 with Hive 0.11. I have succesfully inserted into >> hive/hdfs some csv-files in seperate tables. selects and joins work >> flawlessly. When trying to analyse some data, i needed to make use of the >> built in functions of hive like: >> >> - substr >> - to_date >> - rand >> - etc. >> >> for example: >> >> select sid, request_id, to_date(times), to_unix_timestamp(times) from >> contents where sid = '5000000032066010373'; >> >> sid and request id are strings here, times is a timestamp column >> Unfortanetely i only get errors (always the same error stack) when using >> these functions: >> >> java.lang.RuntimeException: Error in configuring object >> at >> org.apache.hadoop.util.ReflectionUtils.setJobConf(ReflectionUtils.java:93) >> at >> org.apache.hadoop.util.ReflectionUtils.setConf(ReflectionUtils.java:64) >> at >> org.apache.hadoop.util.ReflectionUtils.newInstance(ReflectionUtils.java:117) >> at >> org.apache.hadoop.mapred.MapTask.runOldMapper(MapTask.java:354) >> at org.apache.hadoop.mapred.MapTask.run(MapTask.java:307) >> at org.apache.hadoop.mapred.Child.main(Child.java:170) >> Caused by: java.lang.reflect.InvocationTargetException >> at sun.reflect.NativeMethodAccessorImpl.invoke0(Native Method) >> at >> sun.reflect.NativeMethodAccessorImpl.invoke(NativeMethodAccessorImpl.java:57) >> at >> sun.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:43) >> at java.lang.reflect.Method.invoke(Method.java:601) >> at >> org.apache.hadoop.util.ReflectionUtils.setJobConf(ReflectionUtils.java:88) >> ... 5 more >> Caused by: java.lang.RuntimeException: Error in configuring object >> at >> org.apache.hadoop.util.ReflectionUtils.setJobConf(ReflectionUtils.java:93) >> at >> org.apache.hadoop.util.ReflectionUtils.setConf(ReflectionUtils.java:64) >> at >> org.apache.hadoop.util.ReflectionUtils.newInstance(ReflectionUtils.java:117) >> at >> org.apache.hadoop.mapred.MapRunner.configure(MapRunner.java:34) >> ... 10 more >> Caused by: java.lang.reflect.InvocationTargetException >> at sun.reflect.NativeMethodAccessorImpl.invoke0(Native Method) >> at >> sun.reflect.NativeMethodAccessorImpl.invoke(NativeMethodAccessorImpl.java:57) >> at >> sun.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:43) >> at java.lang.reflect.Method.invoke(Method.java:601) >> at >> org.apache.hadoop.util.ReflectionUtils.setJobConf(ReflectionUtils.java:88) >> ... 13 more >> Caused by: java.lang.RuntimeException: Map operator initialization >> failed >> at >> org.apache.hadoop.hive.ql.exec.ExecMapper.configure(ExecMapper.java:121) >> ... 18 more >> Caused by: java.lang.NoClassDefFoundError: >> org/codehaus/jackson/JsonFactory >> at >> org.apache.hadoop.hive.ql.udf.generic.GenericUDTFJSONTuple.<clinit>(GenericUDTFJSONTuple.java:56) >> at >> sun.reflect.NativeConstructorAccessorImpl.newInstance0(Native Method) >> at >> sun.reflect.NativeConstructorAccessorImpl.newInstance(NativeConstructorAccessorImpl.java:57) >> at >> sun.reflect.DelegatingConstructorAccessorImpl.newInstance(DelegatingConstructorAccessorImpl.java:45) >> at >> java.lang.reflect.Constructor.newInstance(Constructor.java:525) >> at >> org.apache.hadoop.util.ReflectionUtils.newInstance(ReflectionUtils.java:113) >> at >> org.apache.hadoop.hive.ql.exec.FunctionRegistry.registerGenericUDTF(FunctionRegistry.java:526) >> at >> org.apache.hadoop.hive.ql.exec.FunctionRegistry.registerGenericUDTF(FunctionRegistry.java:520) >> at >> org.apache.hadoop.hive.ql.exec.FunctionRegistry.<clinit>(FunctionRegistry.java:423) >> at >> org.apache.hadoop.hive.ql.exec.DefaultUDFMethodResolver.getEvalMethod(DefaultUDFMethodResolver.java:59) >> at >> org.apache.hadoop.hive.ql.udf.generic.GenericUDFBridge.initialize(GenericUDFBridge.java:154) >> at >> org.apache.hadoop.hive.ql.udf.generic.GenericUDF.initializeAndFoldConstants(GenericUDF.java:111) >> at >> org.apache.hadoop.hive.ql.exec.ExprNodeGenericFuncEvaluator.initialize(ExprNodeGenericFuncEvaluator.java:141) >> at >> org.apache.hadoop.hive.ql.exec.Operator.initEvaluators(Operator.java:970) >> at >> org.apache.hadoop.hive.ql.exec.Operator.initEvaluatorsAndReturnStruct(Operator.java:996) >> at >> org.apache.hadoop.hive.ql.exec.SelectOperator.initializeOp(SelectOperator.java:60) >> at >> org.apache.hadoop.hive.ql.exec.Operator.initialize(Operator.java:375) >> at >> org.apache.hadoop.hive.ql.exec.Operator.initialize(Operator.java:451) >> at >> org.apache.hadoop.hive.ql.exec.Operator.initializeChildren(Operator.java:407) >> at >> org.apache.hadoop.hive.ql.exec.FilterOperator.initializeOp(FilterOperator.java:78) >> at >> org.apache.hadoop.hive.ql.exec.Operator.initialize(Operator.java:375) >> at >> org.apache.hadoop.hive.ql.exec.Operator.initialize(Operator.java:451) >> at >> org.apache.hadoop.hive.ql.exec.Operator.initializeChildren(Operator.java:407) >> at >> org.apache.hadoop.hive.ql.exec.TableScanOperator.initializeOp(TableScanOperator.java:186) >> at >> org.apache.hadoop.hive.ql.exec.Operator.initialize(Operator.java:375) >> at >> org.apache.hadoop.hive.ql.exec.MapOperator.initializeOp(MapOperator.java:543) >> at >> org.apache.hadoop.hive.ql.exec.Operator.initialize(Operator.java:375) >> at >> org.apache.hadoop.hive.ql.exec.ExecMapper.configure(ExecMapper.java:100) >> ... 18 more >> Caused by: java.lang.ClassNotFoundException: >> org.codehaus.jackson.JsonFactory >> >> what am i doing wrong here? the jackson-core-asl-1.8.8.jar is in the >> $HIVE_HOME/lib directory ... >> >> SHOW FUNCTIONS; >> >> shows me that these functions are in there ... i already tried >> downgrading to hive 0.10 but the error is the same over there. i need to >> work with hadoop 0.20, so unfortunately i can't try hadoop 1.x.x >> >> thanks in advance >> cheers >> Wolli >> > >