[ 
https://issues.apache.org/jira/browse/SQOOP-1393?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=14092544#comment-14092544
 ] 

Richard commented on SQOOP-1393:
--------------------------------

There are 2 solutions:
1. Use Kite to generate hive table, which is similar with function of import 
into hdfs as Parquet file (SQOOP-1390). Set the Kite dataset uri like 
"dataset:hive?dataset=test&hdfs:host=server-360&hdfs:port:8020" only to 
generate hive table directly.
2. use Sqoop to generate hive table, which is similar with function of import 
into hive as text file. There are 3 steps. First, import into hdfs as Parquet 
file (SQOOP-1390). Second, use Sqoop to create table in hive, like text file. 
Third, move Parquet file to hive warehouse (/user/hive/warehouse).


> Import data from database to Hive as Parquet files
> --------------------------------------------------
>
>                 Key: SQOOP-1393
>                 URL: https://issues.apache.org/jira/browse/SQOOP-1393
>             Project: Sqoop
>          Issue Type: Sub-task
>          Components: tools
>            Reporter: Qian Xu
>            Assignee: Richard
>
> Import data to Hive as Parquet file can be separated into two steps:
> 1. Import an individual table from an RDBMS to HDFS as a set of Parquet files.
> 2. Import the data into Hive by generating and executing a CREATE TABLE 
> statement to define the data's layout in Hive with Parquet format table



--
This message was sent by Atlassian JIRA
(v6.2#6252)

Reply via email to