I have copy paste the ROW in office writer where i saw its # separated... yeah \N values representing NULL.. the version of sqoop is *Sqoop 1.4.2 git commit id Compiled by ag on Tue Aug 14 17:37:19 IST 2012*
On Tue, Jun 18, 2013 at 5:01 PM, Nitin Pawar <nitinpawar...@gmail.com>wrote: > is "#" your field separator? > also the separator is normally an octal representation so you can give it > a try. > > why does your columns have \N as values? is it for NULL ? > > what version of sqoop are you using? > > > On Tue, Jun 18, 2013 at 5:00 PM, Hamza Asad <hamza.asa...@gmail.com>wrote: > >> im executing following command* >> sqoop export --connect jdbc:mysql://localhost/xxxx --table >> dump_hive_events_details --export-dir hive/warehouse/xxxx.db/events_details >> --input-null-non-string \N --input-fields-terminated-by '#' --username >> xxxxxxxx --password xxxxxxxxx* >> * >> 13/06/18 16:26:44 INFO mapred.JobClient: Task Id : >> attempt_201306170658_0106_m_000001_0, Status : FAILED >> java.lang.NumberFormatException: For input string: "8119844 1 4472499 >> 2013-01-29 00:00:00.0 1 4 1 \N \N \N \N \N \N \N \N \N \N 8 \N \N \N \N \N >> 1 \N \N 3 2 \N 1" >> at >> java.lang.NumberFormatException.forInputString(NumberFormatException.java:65) >> at java.lang.Integer.parseInt(Integer.java:492) >> at java.lang.Integer.valueOf(Integer.java:582) >> at >> dump_hive_events_details.__loadFromFields(dump_hive_events_details.java:949) >> at dump_hive_events_details.parse(dump_hive_events_details.java:901) >> at >> org.apache.sqoop.mapreduce.TextExportMapper.map(TextExportMapper.java:77) >> at >> org.apache.sqoop.mapreduce.TextExportMapper.map(TextExportMapper.java:36) >> at org.apache.hadoop.mapreduce.Mapper.run(Mapper.java:144) >> at >> org.apache.sqoop.mapreduce.AutoProgressMapper.run(AutoProgressMapper.java:182) >> at org.apache.hadoop.mapred.MapTask.runNewMapper(MapTask.java:764) >> at org.apache.hadoop.mapred.MapTask.run(MapTask.java:370) >> at org.apache.hadoop.mapred.Child$4.run(Child.java:255) >> at java.security.AccessController.doPrivileged(Native Method) >> at javax.security.auth.Subject.doAs(Subject.java:415) >> at >> org.apache.hadoop.security.UserGroupInformation.doAs(UserGroupInformation.java:1121) >> at org.apache.hadoop.mapred.Child.main(Child.java:249) >> * >> >> >> On Tue, Jun 18, 2013 at 4:07 PM, Nitin Pawar <nitinpawar...@gmail.com>wrote: >> >>> check the option --input-fields-terminated-by in sqoop export >>> >>> >>> On Tue, Jun 18, 2013 at 4:31 PM, Hamza Asad <hamza.asa...@gmail.com>wrote: >>> >>>> I want to export my table in mysql and for that i'm using sqoop export >>>> command but in HDFS i've data apparantly without any field seperator But it >>>> does contain some field separator. data is saved in the format as shown >>>> below >>>> *8119844 1 4472499 2013-01-29 00:00:00.0 1 4 1 \N \N \N \N \N \N \N \N >>>> \N \N 8 \N \N \N \N \N 1 \N \N 3 2 \N 1* >>>> how can i export this type of data to mysql and what field separator i >>>> mention it there.. Please help >>>> >>>> -- >>>> *Muhammad Hamza Asad* >>>> >>> >>> >>> >>> -- >>> Nitin Pawar >>> >> >> >> >> -- >> *Muhammad Hamza Asad* >> > > > > -- > Nitin Pawar > -- *Muhammad Hamza Asad*