never mind, replacing schema.literal with 'avro.schema.literal' worked for
us!!


On Mon, Dec 16, 2013 at 4:01 PM, Sagar Mehta <sagarme...@gmail.com> wrote:

> Hi Guys,
>
> We have an avro backed production table bucketed by a column and it works
> fine with the old Hive = 0.7 on chd3u2.
>
> Now we have moved to Hive 0.12 on HDP 2.0 and are getting bucketing issues
> with this table.
>
> Below is the create table statement
>
> CREATE TABLE test_bucketing
>   PARTITIONED BY (year INT, month INT, day INT, hour INT)
>   CLUSTERED BY (akey) INTO 32 BUCKETS
>   ROW FORMAT SERDE 'org.apache.hadoop.hive.serde2.avro.AvroSerDe'
>   WITH SERDEPROPERTIES (
>     'schema.literal'='{
>   "type" : "record",
>   "name" : "NgEventETLSimple",
>   "namespace" : "com.ngmoco.ngpipes.etl.events",
>   "fields" : [ {
>     "name" : "akey",
>     "type" : "string"
>   } ]
> }')
>   STORED AS INPUTFORMAT
>   'org.apache.hadoop.hive.ql.io.avro.AvroContainerInputFormat'
>   OUTPUTFORMAT
>   'org.apache.hadoop.hive.ql.io.avro.AvroContainerOutputFormat'
>
> and this is the error we are seeing
>
> FAILED: Execution Error, return code 1 from
> org.apache.hadoop.hive.ql.exec.DDLTask. Bucket columns akey is not part of
> the table columns
> ([FieldSchema(name:error_error_error_error_error_error_error, type:string,
> comment:from deserializer), FieldSchema(name:cannot_determine_schema,
> type:string, comment:from deserializer), FieldSchema(name:check,
> type:string, comment:from deserializer), FieldSchema(name:schema,
> type:string, comment:from deserializer), FieldSchema(name:url, type:string,
> comment:from deserializer), FieldSchema(name:and, type:string, comment:from
> deserializer), FieldSchema(name:literal, type:string, comment:from
> deserializer)]
>
> As you can see the akey is mentioned as a part of the table definition,
> however it still complains that akey is not part of the table definition.
>
> I can confirm that independently the avro serde works fine [if we drop the
> bucketing clause] and independently bucketing works fine on a test table
> [without avro serde]. *However together bucketing for an avro backed
> table is giving us issues on Hive 0.12*
>
> Any idea what is going on? Any help/pointers would be useful.
>
> Sagar
> PS - The actual table in production has way more columns but I don't think
> that is relevant for the issue at hand here.
>

Reply via email to