Nitin,
       Issue is not with the INT or BIGINT (as i have verified both),
exception is same.. Issue is with some thing else.. Please sort out any
solution... following exception still raising (# in input string is not
visible in terminal and is translated to # when copied to office writer
which i pasted below)
*java.lang.NumberFormatException: For input string: "** *433649#1#534782#2"
***    at
java.lang.NumberFormatException.forInputString(NumberFormatException.java:65)
    at java.lang.Long.parseLong(Long.java:441)
    at java.lang.Long.valueOf(Long.java:540)
    at
dump_hive_events_details.__loadFromFields(dump_hive_events_details.java:949)
    at dump_hive_events_details.parse(dump_hive_events_details.java:901)
    at
org.apache.sqoop.mapreduce.TextExportMapper.map(TextExportMapper.java:77)
    at
org.apache.sqoop.mapreduce.TextExportMapper.map(TextExportMapper.java:36)
    at org.apache.hadoop.mapreduce.Mapper.run(Mapper.java:144)
    at
org.apache.sqoop.mapreduce.AutoProgressMapper.run(AutoProgressMapper.java:182)
    at org.apache.hadoop.mapred.MapTask.runNewMapper(MapTask.java:764)
    at org.apache.hadoop.mapred.MapTask.run(MapTask.java:370)
    at org.apache.hadoop.mapred.Child$4.run(Child.java:255)
    at java.lang.Long.parseLong(Long.java:441)
    at java.lang.Long.valueOf(Long.java:540)
    at
dump_hive_events_details.__loadFromFields(dump_hive_events_details.java:949)
    at dump_hive_events_details.parse(dump_hive_events_details.java:901)
    at
org.apache.sqoop.mapreduce.TextExportMapper.map(TextExportMapper.java:77)
    at
org.apache.sqoop.mapreduce.TextExportMapper.map(TextExportMapper.java:36)
    at org.apache.hadoop.mapreduce.Mapper.run(Mapper.java:144)
    at
org.apache.sqoop.mapreduce.AutoProgressMapper.run(AutoProgressMapper.java:182)
    at org.apache.hadoop.mapred.MapTask.runNewMapper(MapTask.java:764)
    at org.apache.hadoop.mapred.MapTask.run(MapTask.java:370)
    at org.apache.hadoop.mapred.Child$4.run(Child.java:255)
    at java.security.AccessController.doPrivileged(Native Method)
    at javax.security.auth.Subject.doAs(Subject.java:415)
    at
org.apache.hadoop.security.UserGroupInformation.doAs(UserGroupInformation.java:1121)
    at org.apache.hadoop.mapred.Child.main(Child.java:249)
*


On Tue, Jun 18, 2013 at 5:53 PM, Nitin Pawar <nitinpawar...@gmail.com>wrote:

> can you change your mysql schema to have bigint instead of just int.
> for more you can refer this
> http://stackoverflow.com/questions/16886668/why-sqoop-fails-on-numberformatexception-for-numeric-column-during-the-export-fr
>
>
> On Tue, Jun 18, 2013 at 5:52 PM, Hamza Asad <hamza.asa...@gmail.com>wrote:
>
>> Attached are the schema files of both HIVE and mySql tables
>>
>>
>> On Tue, Jun 18, 2013 at 5:11 PM, Nitin Pawar <nitinpawar...@gmail.com>wrote:
>>
>>>  for the number format exception, can you share your mysql schema (put
>>> as attachment and not inline in mail). If you have created table with int
>>> .. try to switch the column with bigint
>>>
>>>
>>>
>>> On Tue, Jun 18, 2013 at 5:37 PM, Hamza Asad <hamza.asa...@gmail.com>wrote:
>>>
>>>> I have copy paste the ROW in office writer where i saw its #
>>>> separated...
>>>> yeah \N values representing NULL..
>>>> the version of sqoop is
>>>> *Sqoop 1.4.2
>>>> git commit id
>>>> Compiled by ag on Tue Aug 14 17:37:19 IST 2012*
>>>>
>>>>
>>>> On Tue, Jun 18, 2013 at 5:01 PM, Nitin Pawar 
>>>> <nitinpawar...@gmail.com>wrote:
>>>>
>>>>> is "#" your field separator?
>>>>> also the separator is normally an octal representation so you can give
>>>>> it a try.
>>>>>
>>>>> why does your columns have \N as values? is it for NULL ?
>>>>>
>>>>> what version of sqoop are you using?
>>>>>
>>>>>
>>>>> On Tue, Jun 18, 2013 at 5:00 PM, Hamza Asad <hamza.asa...@gmail.com>wrote:
>>>>>
>>>>>> im executing following command*
>>>>>> sqoop export --connect jdbc:mysql://localhost/xxxx --table
>>>>>> dump_hive_events_details --export-dir 
>>>>>> hive/warehouse/xxxx.db/events_details
>>>>>> --input-null-non-string \N --input-fields-terminated-by '#' --username
>>>>>> xxxxxxxx --password xxxxxxxxx*
>>>>>> *
>>>>>> 13/06/18 16:26:44 INFO mapred.JobClient: Task Id :
>>>>>> attempt_201306170658_0106_m_000001_0, Status : FAILED
>>>>>> java.lang.NumberFormatException: For input string: "8119844 1 4472499
>>>>>> 2013-01-29 00:00:00.0 1 4 1 \N \N \N \N \N \N \N \N \N \N 8 \N \N \N \N 
>>>>>> \N
>>>>>> 1 \N \N 3 2 \N 1"
>>>>>>     at
>>>>>> java.lang.NumberFormatException.forInputString(NumberFormatException.java:65)
>>>>>>     at java.lang.Integer.parseInt(Integer.java:492)
>>>>>>     at java.lang.Integer.valueOf(Integer.java:582)
>>>>>>     at
>>>>>> dump_hive_events_details.__loadFromFields(dump_hive_events_details.java:949)
>>>>>>     at
>>>>>> dump_hive_events_details.parse(dump_hive_events_details.java:901)
>>>>>>     at
>>>>>> org.apache.sqoop.mapreduce.TextExportMapper.map(TextExportMapper.java:77)
>>>>>>     at
>>>>>> org.apache.sqoop.mapreduce.TextExportMapper.map(TextExportMapper.java:36)
>>>>>>     at org.apache.hadoop.mapreduce.Mapper.run(Mapper.java:144)
>>>>>>     at
>>>>>> org.apache.sqoop.mapreduce.AutoProgressMapper.run(AutoProgressMapper.java:182)
>>>>>>     at org.apache.hadoop.mapred.MapTask.runNewMapper(MapTask.java:764)
>>>>>>     at org.apache.hadoop.mapred.MapTask.run(MapTask.java:370)
>>>>>>     at org.apache.hadoop.mapred.Child$4.run(Child.java:255)
>>>>>>     at java.security.AccessController.doPrivileged(Native Method)
>>>>>>     at javax.security.auth.Subject.doAs(Subject.java:415)
>>>>>>     at
>>>>>> org.apache.hadoop.security.UserGroupInformation.doAs(UserGroupInformation.java:1121)
>>>>>>     at org.apache.hadoop.mapred.Child.main(Child.java:249)
>>>>>> *
>>>>>>
>>>>>>
>>>>>> On Tue, Jun 18, 2013 at 4:07 PM, Nitin Pawar <nitinpawar...@gmail.com
>>>>>> > wrote:
>>>>>>
>>>>>>> check the option --input-fields-terminated-by in sqoop export
>>>>>>>
>>>>>>>
>>>>>>> On Tue, Jun 18, 2013 at 4:31 PM, Hamza Asad 
>>>>>>> <hamza.asa...@gmail.com>wrote:
>>>>>>>
>>>>>>>> I want to export my table in mysql and for that i'm using sqoop
>>>>>>>> export command but in HDFS i've data apparantly without any field 
>>>>>>>> seperator
>>>>>>>> But it does contain some field separator. data is saved in the format 
>>>>>>>> as
>>>>>>>> shown below
>>>>>>>> *8119844 1 4472499 2013-01-29 00:00:00.0 1 4 1 \N \N \N \N \N \N
>>>>>>>> \N \N \N \N 8 \N \N \N \N \N 1 \N \N 3 2 \N 1*
>>>>>>>> how can i export this type of data to mysql and what field
>>>>>>>> separator i mention it there.. Please help
>>>>>>>>
>>>>>>>> --
>>>>>>>> *Muhammad Hamza Asad*
>>>>>>>>
>>>>>>>
>>>>>>>
>>>>>>>
>>>>>>> --
>>>>>>> Nitin Pawar
>>>>>>>
>>>>>>
>>>>>>
>>>>>>
>>>>>> --
>>>>>> *Muhammad Hamza Asad*
>>>>>>
>>>>>
>>>>>
>>>>>
>>>>> --
>>>>> Nitin Pawar
>>>>>
>>>>
>>>>
>>>>
>>>> --
>>>> *Muhammad Hamza Asad*
>>>>
>>>
>>>
>>>
>>> --
>>> Nitin Pawar
>>>
>>
>>
>>
>> --
>> *Muhammad Hamza Asad*
>>
>
>
>
> --
> Nitin Pawar
>



-- 
*Muhammad Hamza Asad*

Reply via email to