Which version of Hive are you on? Hive 2.1 should automatically handle the type conversions from the file to the table.
.. Owen On Wed, Dec 14, 2016 at 9:36 AM, Daniel Haviv < daniel.ha...@veracity-group.com> wrote: > Hi, > I have an ORC table where one of the fields was an int and is now a bigint. > Whenever I query a partition before the schema change I encounter the > following error: > Error: java.io.IOException: java.io.IOException: > java.lang.ClassCastException: org.apache.hadoop.io.IntWritable cannot be > cast to org.apache.hadoop.io.LongWritable > at org.apache.hadoop.hive.io.HiveIOExceptionHandlerChain. > handleRecordReaderNextException(HiveIOExceptionHandlerChain.java:121) > at org.apache.hadoop.hive.io.HiveIOExceptionHandlerUtil. > handleRecordReaderNextException(HiveIOExceptionHandlerUtil.java:77) > at org.apache.hadoop.hive.shims.HadoopShimsSecure$ > CombineFileRecordReader.doNextWithExceptionHandler( > HadoopShimsSecure.java:226) > at org.apache.hadoop.hive.shims.HadoopShimsSecure$ > CombineFileRecordReader.next(HadoopShimsSecure.java:136) > at org.apache.hadoop.mapred.MapTask$TrackedRecordReader. > moveToNext(MapTask.java:199) > at org.apache.hadoop.mapred.MapTask$TrackedRecordReader. > next(MapTask.java:185) > at org.apache.hadoop.mapred.MapRunner.run(MapRunner.java:52) > at org.apache.hadoop.mapred.MapTask.runOldMapper(MapTask.java:453) > at org.apache.hadoop.mapred.MapTask.run(MapTask.java:343) > at org.apache.hadoop.mapred.YarnChild$2.run(YarnChild.java:164) > at java.security.AccessController.doPrivileged(Native Method) > at javax.security.auth.Subject.doAs(Subject.java:415) > at org.apache.hadoop.security.UserGroupInformation.doAs( > UserGroupInformation.java:1698) > > I tried to manually go through the old partitions and set that column to > int but I'm still getting the same exceptions. > I expected promoting an int to a bigint shouldn't cause any problems. > > Am I doing something wrong ? > > Thank you, > Daniel >