[ https://issues.apache.org/jira/browse/SQOOP-1393?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=14092544#comment-14092544 ]
Richard commented on SQOOP-1393: -------------------------------- There are 2 solutions: 1. Use Kite to generate hive table, which is similar with function of import into hdfs as Parquet file (SQOOP-1390). Set the Kite dataset uri like "dataset:hive?dataset=test&hdfs:host=server-360&hdfs:port:8020" only to generate hive table directly. 2. use Sqoop to generate hive table, which is similar with function of import into hive as text file. There are 3 steps. First, import into hdfs as Parquet file (SQOOP-1390). Second, use Sqoop to create table in hive, like text file. Third, move Parquet file to hive warehouse (/user/hive/warehouse). > Import data from database to Hive as Parquet files > -------------------------------------------------- > > Key: SQOOP-1393 > URL: https://issues.apache.org/jira/browse/SQOOP-1393 > Project: Sqoop > Issue Type: Sub-task > Components: tools > Reporter: Qian Xu > Assignee: Richard > > Import data to Hive as Parquet file can be separated into two steps: > 1. Import an individual table from an RDBMS to HDFS as a set of Parquet files. > 2. Import the data into Hive by generating and executing a CREATE TABLE > statement to define the data's layout in Hive with Parquet format table -- This message was sent by Atlassian JIRA (v6.2#6252)