On Fri, Jun 2, 2017 at 12:07 PM, Nishanth S <nishanth.2...@gmail.com> wrote:

> Hello hive users,
>
> We are looking at migrating  files(less than 5 Mb of data in total) with
> variable record lengths from a mainframe system to hive.You could think of
> this as metadata.Each of these records can have columns  ranging from 3 to
>  n( means  each record type have different number of columns) based on
> record type.What would be the best strategy to migrate this  to hive .I was
> thinking of converting these files  into one  variable length csv file and
> then importing them to a hive table .Hive table will consist of 4 columns
> with the 4th column having comma separated list of  values from column
> column 4 to n.Are there other alternative or better approaches for this
> solution.Appreciate any  feedback on this.
>
> Thanks,
> Nishanth
>

Hive supports complex types like List, Map, and Struct and they can be
arbitrarily nested. If the nested data has a schema that may be your best
option. Potentially using thrift/avro/parquet/protobuf support.

Otherwise you can store the data as Json and at read time parse things out
using json udfs.

Edward

Reply via email to