Vinodh Kumar R created SQOOP-3482:
-------------------------------------

             Summary: sqoop-import fails when importing records with datatype 
'Money' in PostgreSQL
                 Key: SQOOP-3482
                 URL: https://issues.apache.org/jira/browse/SQOOP-3482
             Project: Sqoop
          Issue Type: Bug
          Components: connectors/postgresql
    Affects Versions: 1.4.7
         Environment: +*!1.JPG!!2.JPG!*+
            Reporter: Vinodh Kumar R
         Attachments: 1.JPG, 2.JPG

ssue description :
We are trying to sqoop-import records from PostgreSQL server, which contain a 
fields datatype 'money'.
During the course of import the Postgres JDBC driver is trying read this field 
as datatype 'double' and eventually it fails to import the record and throws 
the below exception -
Bad value for type double : 100,000.00

Steps to reproduce the behaviour:

Environment details -
Tried with Hadoop 2.7.1 and 3.1.4
sqoop-version = 1.4.6 and 1.4.7
Postgre JDBC driver = postgresql-42.2.16.jar
Postgre SQL 12.4
Linux OS - Ubuntu 20.04.1 LTS

Database Name : stg_db
Table Name : tbl_cust
Columns and DataType :
cust_f_name varchar(25),
cust_l_name varchar(25),
trans_amount money
Table Contains 1 row as below -
insert into tbl_cust values ('VR','Kumar',100000)

Sqoop Statement (Connection string)
sqoop-import --connect jdbc:postgresql://192.168.125.130:5432/stg_db
--username postgres --password postgres
--table tbl_cust -m 1 --target-dir tbl_cust --delete-target-dir;

Below is the log message -
_20/09/08 14:24:23 INFO mapreduce.Job: Task Id : 
attempt_1599505663642_0011_m_000000_0, Status : FAILED_
_Error: java.io.IOException: SQLException in nextKeyValue_
_at 
org.apache.sqoop.mapreduce.db.DBRecordReader.nextKeyValue(DBRecordReader.java:277)_
_at 
org.apache.hadoop.mapred.MapTask$NewTrackingRecordReader.nextKeyValue(MapTask.java:556)_
_at 
org.apache.hadoop.mapreduce.task.MapContextImpl.nextKeyValue(MapContextImpl.java:80)_
_at 
org.apache.hadoop.mapreduce.lib.map.WrappedMapper$Context.nextKeyValue(WrappedMapper.java:91)_
_at org.apache.hadoop.mapreduce.Mapper.run(Mapper.java:145)_
_at 
org.apache.sqoop.mapreduce.AutoProgressMapper.run(AutoProgressMapper.java:64)_
_at org.apache.hadoop.mapred.MapTask.runNewMapper(MapTask.java:787)_
_at org.apache.hadoop.mapred.MapTask.run(MapTask.java:341)_
_at org.apache.hadoop.mapred.YarnChild$2.run(YarnChild.java:164)_
_at java.security.AccessController.doPrivileged(Native Method)_
_at javax.security.auth.Subject.doAs(Subject.java:422)_
_at 
org.apache.hadoop.security.UserGroupInformation.doAs(UserGroupInformation.java:1657)_
_at org.apache.hadoop.mapred.YarnChild.main(YarnChild.java:158)_
*_Caused by: org.postgresql.util.PSQLException: Bad value for type double : 
100,000.00_*
_at org.postgresql.jdbc.PgResultSet.toDouble(PgResultSet.java:3104)_
_at org.postgresql.jdbc.PgResultSet.getDouble(PgResultSet.java:2432)_
_at 
org.apache.sqoop.lib.JdbcWritableBridge.readDouble(JdbcWritableBridge.java:86)_
_at 
com.cloudera.sqoop.lib.JdbcWritableBridge.readDouble(JdbcWritableBridge.java:69)_
_at tabletodrop.readFields(tabletodrop.java:106)_
_at 
org.apache.sqoop.mapreduce.db.DBRecordReader.nextKeyValue(DBRecordReader.java:244)_
_... 12 more_

However, other tables which doesn't have any field with datatype 'money' gets 
imported without any issues.
Attached screenshot of the record in the postgresql table and the log details 
for reference.



--
This message was sent by Atlassian Jira
(v8.3.4#803005)

Reply via email to