>14/09/12 20:16:55 INFO mapred.JobClient: Task Id :
attempt_201409022012_0543_m_000000_2, Status : FAILED
java.lang.RuntimeException: Should never be used

This is because getRecordWriter is explicitly disabled in Parquet.   That
is why I mentioned in another thread that while HCatalog support in sqoop
is storage agnostic and should in theory support all hive serde's, some
storage formats may not work.

That said, to import into parquet files, you can use the --as-parquetfile
option that was recently introduced.

Thanks

Venkat

On Fri, Sep 12, 2014 at 5:29 PM, Pratik Khadloya (JIRA) <j...@apache.org>
wrote:

>
>     [
> https://issues.apache.org/jira/browse/SQOOP-1393?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=14132376#comment-14132376
> ]
>
> Pratik Khadloya commented on SQOOP-1393:
> ----------------------------------------
>
> I got past the metadata error by running the command using hcatalog.
>
> {code}
> bin/sqoop import -jt myjt:xxxx --connect jdbc:mysql://mydbserver.net/mydb
> --username myuser --password mypwd --query "SELECT ... WHERE \$CONDITIONS"
> --num-mappers 1 --hcatalog-storage-stanza "STORED AS PARQUET"
> --create-hcatalog-table --hcatalog-table abc2
> {code}
>
> But, since i am using hive 0.13, i get the following error which states
> that one should not use MapredParquetOutputFormat with hive 0.13 as it has
> native support for PARQUET files.
> {code}
> 14/09/12 20:16:55 INFO mapred.JobClient: Task Id :
> attempt_201409022012_0543_m_000000_2, Status : FAILED
> java.lang.RuntimeException: Should never be used
>         at
> org.apache.hadoop.hive.ql.io.parquet.MapredParquetOutputFormat.getRecordWriter(MapredParquetOutputFormat.java:77)
>         at
> org.apache.hive.hcatalog.mapreduce.FileOutputFormatContainer.getRecordWriter(FileOutputFormatContainer.java:103)
>         at
> org.apache.hive.hcatalog.mapreduce.HCatOutputFormat.getRecordWriter(HCatOutputFormat.java:260)
>         at
> org.apache.hadoop.mapred.MapTask$NewDirectOutputCollector.<init>(MapTask.java:548)
>         at org.apache.hadoop.mapred.MapTask.runNewMapper(MapTask.java:653)
>         at org.apache.hadoop.mapred.MapTask.run(MapTask.java:330)
>         at org.apache.hadoop.mapred.Child$4.run(Child.java:268)
>         at java.security.AccessController.doPrivileged(Native Method)
>         at javax.security.auth.Subject.doAs(Subject.java:396)
>         at
> org.apache.hadoop.security.UserGroupInformation.doAs(UserGroupInformation.java:1408)
>         at org.apache.hadoop.mapred.Child.main(Child.java:262)
> {code}
>
> Is there any code change planned for supporting hive 0.13 ?
>
> > Import data from database to Hive as Parquet files
> > --------------------------------------------------
> >
> >                 Key: SQOOP-1393
> >                 URL: https://issues.apache.org/jira/browse/SQOOP-1393
> >             Project: Sqoop
> >          Issue Type: Sub-task
> >          Components: tools
> >            Reporter: Qian Xu
> >            Assignee: Richard
> >             Fix For: 1.4.6
> >
> >         Attachments: patch.diff, patch_v2.diff, patch_v3.diff
> >
> >
> > Import data to Hive as Parquet file can be separated into two steps:
> > 1. Import an individual table from an RDBMS to HDFS as a set of Parquet
> files.
> > 2. Import the data into Hive by generating and executing a CREATE TABLE
> statement to define the data's layout in Hive with Parquet format table
>
>
>
> --
> This message was sent by Atlassian JIRA
> (v6.3.4#6332)
>



-- 
Regards

Venkat

Reply via email to