[ https://issues.apache.org/jira/browse/HIVE-2950?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=13405332#comment-13405332 ]
Travis Crawford commented on HIVE-2950: --------------------------------------- In some cases (storing thrift/protobufs) reporting the read-time schema is preferable to the write-time schema. For example, let's say you're storing thrift structs and add a new optional field with default value. In that case all your old records would be automatically upgraded if using the read-time field reporting. I played around with this over the weekend and think it could look something like this. If the partition storage descriptor has a serde that you should get the cols from, then do that. Otherwise, use the fields stored in the metastore. Since this is per-partition you could have some partitions using serde-reported fields, and other partitions using hard-coded ones. {code:title=ql/src/java/org/apache/hadoop/hive/ql/metadata/Partition.java} - public List<FieldSchema> getCols() { - return tPartition.getSd().getCols(); + public List<FieldSchema> getCols() throws HiveException { + if (SerDeUtils.shouldGetColsFromSerDe( + tPartition.getSd().getSerdeInfo().getSerializationLib())) { + return Hive.getFieldsFromDeserializer(table.getTableName(), getDeserializer()); + } else { + return tPartition.getSd().getCols(); + } {code} Thoughts? I still think the original patch (copy table columns into partition storage descriptor at write time) is an improvement, but this dynamic approach would be awesome. Some of our schemas are pretty large so this would save a lot of metastore space :) > Hive should store the full table schema in partition storage descriptors > ------------------------------------------------------------------------ > > Key: HIVE-2950 > URL: https://issues.apache.org/jira/browse/HIVE-2950 > Project: Hive > Issue Type: Bug > Reporter: Travis Crawford > Assignee: Travis Crawford > Attachments: HIVE-2950.D2769.1.patch > > > Hive tables have a schema, which is copied into the partition storage > descriptor when adding a partition. Currently only columns stored in the > table storage descriptor are copied - columns that are reported by the serde > are not copied. Instead of copying the table storage descriptor columns into > the partition columns, the full table schema should be copied. > DETAILS > This is a little long but is necessary to show 3 things: current behavior > when explicitly listing columns, behavior with HIVE-2941 patched in and serde > reported columns, and finally the behavior with this patch (full table schema > copied into the partition storage descriptor). > Here's an example of what currently happens. Note the following: > * the two manually-defined fields defined for the table are listed in the > table storage descriptor. > * both fields are present in the partition storage descriptor > This works great because users who query for a partition can look at its > storage descriptor and get the schema. > {code} > hive> create external table foo_test (name string, age int) partitioned by > (part_dt string); > hive> describe extended foo_test; > OK > name string > age int > part_dt string > > Detailed Table Information Table(tableName:foo_test, dbName:travis_test, > owner:travis, createTime:1334256062, lastAccessTime:0, retention:0, > sd:StorageDescriptor(cols:[FieldSchema(name:name, type:string, comment:null), > FieldSchema(name:age, type:int, comment:null), FieldSchema(name:part_dt, > type:string, comment:null)], > location:hdfs://foo.com/warehouse/travis_test.db/foo_test, > inputFormat:org.apache.hadoop.mapred.TextInputFormat, > outputFormat:org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat, > compressed:false, numBuckets:-1, serdeInfo:SerDeInfo(name:null, > serializationLib:org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe, > parameters:{serialization.format=1}), bucketCols:[], sortCols:[], > parameters:{}, primaryRegionName:, secondaryRegions:[]), > partitionKeys:[FieldSchema(name:part_dt, type:string, comment:null)], > parameters:{EXTERNAL=TRUE, transient_lastDdlTime=1334256062}, > viewOriginalText:null, viewExpandedText:null, tableType:EXTERNAL_TABLE) > Time taken: 0.082 seconds > hive> alter table foo_test add partition (part_dt = '20120331T000000Z') > location 'hdfs://foo.com/foo/2012/03/31/00'; > hive> describe extended foo_test partition (part_dt = '20120331T000000Z'); > OK > name string > age int > part_dt string > > Detailed Partition Information Partition(values:[20120331T000000Z], > dbName:travis_test, tableName:foo_test, createTime:1334256131, > lastAccessTime:0, sd:StorageDescriptor(cols:[FieldSchema(name:name, > type:string, comment:null), FieldSchema(name:age, type:int, comment:null), > FieldSchema(name:part_dt, type:string, comment:null)], > location:hdfs://foo.com/foo/2012/03/31/00, > inputFormat:org.apache.hadoop.mapred.TextInputFormat, > outputFormat:org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat, > compressed:false, numBuckets:-1, serdeInfo:SerDeInfo(name:null, > serializationLib:org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe, > parameters:{serialization.format=1}), bucketCols:[], sortCols:[], > parameters:{}, primaryRegionName:, secondaryRegions:[]), > parameters:{transient_lastDdlTime=1334256131}) > {code} > CURRENT BEHAVIOR WITH HIVE-2941 PATCHED IN > Now let's examine what happens when creating a table when the serde reports > the schema. Notice the following: > * The table storage descriptor contains an empty list of columns. However, > the table schema is available from the serde reflecting on the serialization > class. > * The partition storage descriptor does contain a single "part_dt" column > that was copied from the table partition keys. The actual data columns are > not present. > {code} > hive> create external table travis_test.person_test partitioned by (part_dt > string) row format serde "com.twitter.elephantbird.hive.serde.ThriftSerDe" > with serdeproperties > ("serialization.class"="com.twitter.elephantbird.examples.thrift.Person") > stored as inputformat > "com.twitter.elephantbird.mapred.input.HiveMultiInputFormat" outputformat > "org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat"; > OK > Time taken: 0.08 seconds > hive> describe extended person_test; > OK > name struct<first_name:string,last_name:string> from deserializer > id int from deserializer > email string from deserializer > phones array<struct<number:string,type:struct<value:int>>> from > deserializer > part_dt string > > Detailed Table Information Table(tableName:person_test, > dbName:travis_test, owner:travis, createTime:1334256942, lastAccessTime:0, > retention:0, sd:StorageDescriptor(cols:[], > location:hdfs://foo.com/warehouse/travis_test.db/person_test, > inputFormat:com.twitter.elephantbird.mapred.input.HiveMultiInputFormat, > outputFormat:org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat, > compressed:false, numBuckets:-1, serdeInfo:SerDeInfo(name:null, > serializationLib:com.twitter.elephantbird.hive.serde.ThriftSerDe, > parameters:{serialization.class=com.twitter.elephantbird.examples.thrift.Person, > serialization.format=1}), bucketCols:[], sortCols:[], parameters:{}, > primaryRegionName:, secondaryRegions:[]), > partitionKeys:[FieldSchema(name:part_dt, type:string, comment:null)], > parameters:{EXTERNAL=TRUE, transient_lastDdlTime=1334256942}, > viewOriginalText:null, viewExpandedText:null, tableType:EXTERNAL_TABLE) > Time taken: 0.147 seconds > hive> alter table person_test add partition (part_dt = '20120331T000000Z') > location 'hdfs://foo.com/foo/2012/03/31/00'; > OK > Time taken: 0.149 seconds > hive> describe extended person_test partition (part_dt = '20120331T000000Z'); > OK > part_dt string > > Detailed Partition Information Partition(values:[20120331T000000Z], > dbName:travis_test, tableName:person_test, createTime:1334257029, > lastAccessTime:0, sd:StorageDescriptor(cols:[FieldSchema(name:part_dt, > type:string, comment:null)], location:hdfs://foo.com/foo/2012/03/31/00, > inputFormat:com.twitter.elephantbird.mapred.input.HiveMultiInputFormat, > outputFormat:org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat, > compressed:false, numBuckets:-1, serdeInfo:SerDeInfo(name:null, > serializationLib:com.twitter.elephantbird.hive.serde.ThriftSerDe, > parameters:{serialization.class=com.twitter.elephantbird.examples.thrift.Person, > serialization.format=1}), bucketCols:[], sortCols:[], parameters:{}, > primaryRegionName:, secondaryRegions:[]), > parameters:{transient_lastDdlTime=1334257029}) > Time taken: 0.106 seconds > hive> > {code} > PROPOSED BEHAVIOR > I believe the correct thing to do is copy the full table schema > (serde-reported columns + partition keys) into the partition storage > descriptor. Notice the following: > * Table storage descriptor does not contain any columns, because they are > reported by the serde. > * Partition storage descriptor now contains both the serde-reported schema, > and full table schema. > {code} > hive> create external table travis_test.person_test partitioned by (part_dt > string) row format serde "com.twitter.elephantbird.hive.serde.ThriftSerDe" > with serdeproperties > ("serialization.class"="com.twitter.elephantbird.examples.thrift.Person") > stored as inputformat > "com.twitter.elephantbird.mapred.input.HiveMultiInputFormat" outputformat > "org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat"; > OK > Time taken: 0.076 seconds > hive> describe extended person_test; > > OK > > name struct<first_name:string,last_name:string> > from deserializer > id int from deserializer > email string from deserializer > phones array<struct<number:string,type:struct<value:int>>> from > deserializer > part_dt string > > Detailed Table Information Table(tableName:person_test, > dbName:travis_test, owner:travis, createTime:1334257489, lastAccessTime:0, > retention:0, sd:StorageDescriptor(cols:[], > location:hdfs://foo.com/warehouse/travis_test.db/person_test, > inputFormat:com.twitter.elephantbird.mapred.input.HiveMultiInputFormat, > outputFormat:org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat, > compressed:false, numBuckets:-1, serdeInfo:SerDeInfo(name:null, > serializationLib:com.twitter.elephantbird.hive.serde.ThriftSerDe, > parameters:{serialization.class=com.twitter.elephantbird.examples.thrift.Person, > serialization.format=1}), bucketCols:[], sortCols:[], parameters:{}, > primaryRegionName:, secondaryRegions:[]), > partitionKeys:[FieldSchema(name:part_dt, type:string, comment:null)], > parameters:{EXTERNAL=TRUE, transient_lastDdlTime=1334257489}, > viewOriginalText:null, viewExpandedText:null, tableType:EXTERNAL_TABLE) > Time taken: 0.155 seconds > hive> alter table person_test add partition (part_dt = '20120331T000000Z') > location 'hdfs://foo.com/foo/2012/03/31/00'; > OK > > Time taken: 0.296 seconds > hive> describe extended person_test partition (part_dt = '20120331T000000Z'); > > OK > > name struct<first_name:string,last_name:string> > from deserializer > id int from deserializer > email string from deserializer > phones array<struct<number:string,type:struct<value:int>>> from > deserializer > part_dt string > > Detailed Partition Information Partition(values:[20120331T000000Z], > dbName:travis_test, tableName:person_test, createTime:1334257504, > lastAccessTime:0, sd:StorageDescriptor(cols:[FieldSchema(name:name, > type:struct<first_name:string,last_name:string>, comment:from deserializer), > FieldSchema(name:id, type:int, comment:from deserializer), > FieldSchema(name:email, type:string, comment:from deserializer), > FieldSchema(name:phones, > type:array<struct<number:string,type:struct<value:int>>>, comment:from > deserializer), FieldSchema(name:part_dt, type:string, comment:null)], > location:hdfs://foo.com/foo/2012/03/31/00, > inputFormat:com.twitter.elephantbird.mapred.input.HiveMultiInputFormat, > outputFormat:org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat, > compressed:false, numBuckets:-1, serdeInfo:SerDeInfo(name:null, > serializationLib:com.twitter.elephantbird.hive.serde.ThriftSerDe, > parameters:{serialization.class=com.twitter.elephantbird.examples.thrift.Person, > serialization.format=1}), bucketCols:[], sortCols:[], parameters:{}, > primaryRegionName:, secondaryRegions:[]), > parameters:{transient_lastDdlTime=1334257504}) > Time taken: 0.133 seconds > hive> > {code} -- This message is automatically generated by JIRA. If you think it was sent incorrectly, please contact your JIRA administrators: https://issues.apache.org/jira/secure/ContactAdministrators!default.jspa For more information on JIRA, see: http://www.atlassian.com/software/jira