[ https://issues.apache.org/jira/browse/HIVE-5504?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel ]
Sushanth Sowmyan updated HIVE-5504: ----------------------------------- Attachment: HIVE-5504.patch Attaching patch. Changed Orc's RecordWriter instantiation to fall back to checking jobConf if the orc-specific properties are not present in the provided table properties. This way, users of OrcOutputFormat other than hive have a way of passing these parameters, such as the compression specification to orc. In addition, changed HCat's FileOutputFormatContainer to have a way of special-casing instantiation. We already had one RCFile-specific instantiation, that is now refactored out to a separate class intended to be a collection of special cases, and added in the ability to copy orc-specific table properties from the TableDesc to job properties (which will then later be copied to jobconf by hive) so as pass on these parameters to orc. > OrcOutputFormat honors compression properties only from within hive > --------------------------------------------------------------------- > > Key: HIVE-5504 > URL: https://issues.apache.org/jira/browse/HIVE-5504 > Project: Hive > Issue Type: Bug > Components: HCatalog > Affects Versions: 0.11.0, 0.12.0 > Reporter: Venkat Ranganathan > Attachments: HIVE-5504.patch > > > When we import data into a HCatalog table created with the following storage > description > .. stored as orc tblproperties ("orc.compress"="SNAPPY") > the resultant orc file still uses the default zlib compression > It looks like HCatOutputFormat is ignoring the tblproperties specified. > show tblproperties shows that the table indeed has the properties properly > saved. > An insert/select into the table has the resulting orc file honor the tbl > property. -- This message was sent by Atlassian JIRA (v6.1.5#6160)