Sorry about that, I'm using the columnar SerDe on both tables. Do you need anything else? I don't have the create tables for them, so I can't give you that particular code.
On Tue, Jan 15, 2013 at 4:46 PM, Mark Grover <grover.markgro...@gmail.com>wrote: > I was more interested in knowing if you were using any particular SerDes. > You don't have to list out the columns, just the skeleton create table > statement should do. > > > On Tue, Jan 15, 2013 at 10:43 AM, Fernando Andrés Doglio Turissini < > fernando.dog...@globant.com> wrote: > >> The "data_table" has around 5k fields, all doubles. >> As for the "age_mean" table, here it is: >> >> hive> desc age_mean; >> OK >> id string >> name string >> age_mean double >> Time taken: 0.127 seconds >> >> Does this help? >> >> Thanks! >> Fernando >> >> On Tue, Jan 15, 2013 at 4:35 PM, Mark Grover <grover.markgro...@gmail.com >> > wrote: >> >>> Fernando, >>> Could you share your table definitions as well please? >>> >>> >>> On Tue, Jan 15, 2013 at 10:31 AM, Fernando Andrés Doglio Turissini < >>> fernando.dog...@globant.com> wrote: >>> >>>> Hello everyone, I'm struggling with an exception I'm getting on a >>>> particular query that's driving me crazy! >>>> >>>> Here is the exception I get: >>>> >>>> java.lang.RuntimeException: >>>> org.apache.hadoop.hive.ql.metadata.HiveException: Hive Runtime Error while >>>> processing writable org.apache.hadoop.hive.serde2.colum >>>> nar.BytesRefArrayWritable@71412b61 >>>> at >>>> org.apache.hadoop.hive.ql.exec.ExecMapper.map(ExecMapper.java:161) >>>> at org.apache.hadoop.mapred.MapRunner.run(MapRunner.java:50) >>>> at >>>> org.apache.hadoop.mapred.MapTask.runOldMapper(MapTask.java:441) >>>> at org.apache.hadoop.mapred.MapTask.run(MapTask.java:377) >>>> at org.apache.hadoop.mapred.Child$4.run(Child.java:255) >>>> at java.security.AccessController.doPrivileged(Native Method) >>>> at javax.security.auth.Subject.doAs(Subject.java:396) >>>> at >>>> org.apache.hadoop.security.UserGroupInformation.doAs(UserGroupInformation.java:1132) >>>> at org.apache.hadoop.mapred.Child.main(Child.java:249) >>>> Caused by: org.apache.hadoop.hive.ql.metadata.HiveException: Hive >>>> Runtime Error while processing writable >>>> org.apache.hadoop.hive.serde2.columnar.BytesRefArray >>>> Writable@71412b61 >>>> at >>>> org.apache.hadoop.hive.ql.exec.MapOperator.process(MapOperator.java:524) >>>> at >>>> org.apache.hadoop.hive.ql.exec.ExecMapper.map(ExecMapper.java:143) >>>> ... 8 more >>>> Caused by: java.lang.ArrayIndexOutOfBoundsException: 2 >>>> at >>>> org.apache.hadoop.hive.ql.exec.MapOperator.process(MapOperator.java:506) >>>> ... 9 more >>>> >>>> >>>> Here is the query I'm running: >>>> >>>> INSERT INTO TABLE variance >>>> SELECT id, collect_set(name)[0], SUM( POW(age - age_mean, 2) ) / >>>> count(1) >>>> FROM age_mean join data_table on (age_mean.id = '01' AND data_table.q1 >>>> = 1) >>>> where age is not null and age_mean is not null GROUP BY id; >>>> >>>> It's probably relevant to mention that I'm doing this on an EMR cluster. >>>> >>>> Any idea what might be causing the exception? >>>> >>>> Thanks! >>>> Fernando >>>> >>> >>> >> >