[ https://issues.apache.org/jira/browse/SQOOP-3482?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel ]
Vinodh Kumar R updated SQOOP-3482: ---------------------------------- Description: +*Issue description:*+ We are trying to sqoop-import records from PostgreSQL server, which contain a fields datatype 'money'. During the course of import the Postgres JDBC driver is trying read this field as datatype 'double' and eventually it fails to import the record and throws the below exception - Bad value for type double : 100,000.00 Steps to reproduce the behaviour: +*Environment details -*+ Tried with Hadoop 2.7.1 and 3.1.4 sqoop-version = 1.4.6 and 1.4.7 Postgre JDBC driver = postgresql-42.2.16.jar Postgre SQL 12.4 Linux OS - Ubuntu 20.04.1 LTS Database Name : stg_db Table Name : tbl_cust Columns and DataType : cust_f_name varchar(25), cust_l_name varchar(25), trans_amount money Table Contains 1 row as below - insert into tbl_cust values ('VR','Kumar',100000) +*Sqoop Statement (Connection string)*+ sqoop-import --connect jdbc:postgresql://192.168.0.130:5432/stg_db --username postgres --password cornerstone --table tbl_cust -m 1 --target-dir tbl_cust --delete-target-dir; Below is the log message - _20/09/08 14:24:23 INFO mapreduce.Job: Task Id : attempt_1599505663642_0011_m_000000_0, Status : FAILED_ _Error: java.io.IOException: SQLException in nextKeyValue_ _at org.apache.sqoop.mapreduce.db.DBRecordReader.nextKeyValue(DBRecordReader.java:277)_ _at org.apache.hadoop.mapred.MapTask$NewTrackingRecordReader.nextKeyValue(MapTask.java:556)_ _at org.apache.hadoop.mapreduce.task.MapContextImpl.nextKeyValue(MapContextImpl.java:80)_ _at org.apache.hadoop.mapreduce.lib.map.WrappedMapper$Context.nextKeyValue(WrappedMapper.java:91)_ _at org.apache.hadoop.mapreduce.Mapper.run(Mapper.java:145)_ _at org.apache.sqoop.mapreduce.AutoProgressMapper.run(AutoProgressMapper.java:64)_ _at org.apache.hadoop.mapred.MapTask.runNewMapper(MapTask.java:787)_ _at org.apache.hadoop.mapred.MapTask.run(MapTask.java:341)_ _at org.apache.hadoop.mapred.YarnChild$2.run(YarnChild.java:164)_ _at java.security.AccessController.doPrivileged(Native Method)_ _at javax.security.auth.Subject.doAs(Subject.java:422)_ _at org.apache.hadoop.security.UserGroupInformation.doAs(UserGroupInformation.java:1657)_ _at org.apache.hadoop.mapred.YarnChild.main(YarnChild.java:158)_ *_Caused by: org.postgresql.util.PSQLException: Bad value for type double : 100,000.00_* _at org.postgresql.jdbc.PgResultSet.toDouble(PgResultSet.java:3104)_ _at org.postgresql.jdbc.PgResultSet.getDouble(PgResultSet.java:2432)_ _at org.apache.sqoop.lib.JdbcWritableBridge.readDouble(JdbcWritableBridge.java:86)_ _at com.cloudera.sqoop.lib.JdbcWritableBridge.readDouble(JdbcWritableBridge.java:69)_ _at tabletodrop.readFields(tabletodrop.java:106)_ _at org.apache.sqoop.mapreduce.db.DBRecordReader.nextKeyValue(DBRecordReader.java:244)_ _... 12 more_ However, other tables which doesn't have any field with datatype 'money' gets imported without any issues. Attached screenshot of the record in the postgresql table and the log details for reference. was: +*Issue description:*+ We are trying to sqoop-import records from PostgreSQL server, which contain a fields datatype 'money'. During the course of import the Postgres JDBC driver is trying read this field as datatype 'double' and eventually it fails to import the record and throws the below exception - Bad value for type double : 100,000.00 Steps to reproduce the behaviour: +*Environment details -*+ Tried with Hadoop 2.7.1 and 3.1.4 sqoop-version = 1.4.6 and 1.4.7 Postgre JDBC driver = postgresql-42.2.16.jar Postgre SQL 12.4 Linux OS - Ubuntu 20.04.1 LTS Database Name : stg_db Table Name : tbl_cust Columns and DataType : cust_f_name varchar(25), cust_l_name varchar(25), trans_amount money Table Contains 1 row as below - insert into tbl_cust values ('VR','Kumar',100000) Sqoop Statement (Connection string) sqoop-import --connect jdbc:postgresql://192.168.0.130:5432/stg_db --username postgres --password cornerstone --table tbl_cust -m 1 --target-dir tbl_cust --delete-target-dir; Below is the log message - _20/09/08 14:24:23 INFO mapreduce.Job: Task Id : attempt_1599505663642_0011_m_000000_0, Status : FAILED_ _Error: java.io.IOException: SQLException in nextKeyValue_ _at org.apache.sqoop.mapreduce.db.DBRecordReader.nextKeyValue(DBRecordReader.java:277)_ _at org.apache.hadoop.mapred.MapTask$NewTrackingRecordReader.nextKeyValue(MapTask.java:556)_ _at org.apache.hadoop.mapreduce.task.MapContextImpl.nextKeyValue(MapContextImpl.java:80)_ _at org.apache.hadoop.mapreduce.lib.map.WrappedMapper$Context.nextKeyValue(WrappedMapper.java:91)_ _at org.apache.hadoop.mapreduce.Mapper.run(Mapper.java:145)_ _at org.apache.sqoop.mapreduce.AutoProgressMapper.run(AutoProgressMapper.java:64)_ _at org.apache.hadoop.mapred.MapTask.runNewMapper(MapTask.java:787)_ _at org.apache.hadoop.mapred.MapTask.run(MapTask.java:341)_ _at org.apache.hadoop.mapred.YarnChild$2.run(YarnChild.java:164)_ _at java.security.AccessController.doPrivileged(Native Method)_ _at javax.security.auth.Subject.doAs(Subject.java:422)_ _at org.apache.hadoop.security.UserGroupInformation.doAs(UserGroupInformation.java:1657)_ _at org.apache.hadoop.mapred.YarnChild.main(YarnChild.java:158)_ *_Caused by: org.postgresql.util.PSQLException: Bad value for type double : 100,000.00_* _at org.postgresql.jdbc.PgResultSet.toDouble(PgResultSet.java:3104)_ _at org.postgresql.jdbc.PgResultSet.getDouble(PgResultSet.java:2432)_ _at org.apache.sqoop.lib.JdbcWritableBridge.readDouble(JdbcWritableBridge.java:86)_ _at com.cloudera.sqoop.lib.JdbcWritableBridge.readDouble(JdbcWritableBridge.java:69)_ _at tabletodrop.readFields(tabletodrop.java:106)_ _at org.apache.sqoop.mapreduce.db.DBRecordReader.nextKeyValue(DBRecordReader.java:244)_ _... 12 more_ However, other tables which doesn't have any field with datatype 'money' gets imported without any issues. Attached screenshot of the record in the postgresql table and the log details for reference. > sqoop-import fails when importing records with datatype 'Money' in PostgreSQL > ----------------------------------------------------------------------------- > > Key: SQOOP-3482 > URL: https://issues.apache.org/jira/browse/SQOOP-3482 > Project: Sqoop > Issue Type: Bug > Components: connectors/postgresql > Affects Versions: 1.4.7 > Environment: +*!1.JPG!!2.JPG!*+ > Reporter: Vinodh Kumar R > Priority: Minor > Attachments: 1.JPG, 2.JPG > > > +*Issue description:*+ > We are trying to sqoop-import records from PostgreSQL server, which contain > a fields datatype 'money'. > During the course of import the Postgres JDBC driver is trying read this > field as datatype 'double' and eventually it fails to import the record and > throws the below exception - > Bad value for type double : 100,000.00 > Steps to reproduce the behaviour: > +*Environment details -*+ > Tried with Hadoop 2.7.1 and 3.1.4 > sqoop-version = 1.4.6 and 1.4.7 > Postgre JDBC driver = postgresql-42.2.16.jar > Postgre SQL 12.4 > Linux OS - Ubuntu 20.04.1 LTS > Database Name : stg_db > Table Name : tbl_cust > Columns and DataType : > cust_f_name varchar(25), > cust_l_name varchar(25), > trans_amount money > Table Contains 1 row as below - > insert into tbl_cust values ('VR','Kumar',100000) > +*Sqoop Statement (Connection string)*+ > sqoop-import --connect jdbc:postgresql://192.168.0.130:5432/stg_db > --username postgres --password cornerstone > --table tbl_cust -m 1 --target-dir tbl_cust --delete-target-dir; > Below is the log message - > _20/09/08 14:24:23 INFO mapreduce.Job: Task Id : > attempt_1599505663642_0011_m_000000_0, Status : FAILED_ > _Error: java.io.IOException: SQLException in nextKeyValue_ > _at > org.apache.sqoop.mapreduce.db.DBRecordReader.nextKeyValue(DBRecordReader.java:277)_ > _at > org.apache.hadoop.mapred.MapTask$NewTrackingRecordReader.nextKeyValue(MapTask.java:556)_ > _at > org.apache.hadoop.mapreduce.task.MapContextImpl.nextKeyValue(MapContextImpl.java:80)_ > _at > org.apache.hadoop.mapreduce.lib.map.WrappedMapper$Context.nextKeyValue(WrappedMapper.java:91)_ > _at org.apache.hadoop.mapreduce.Mapper.run(Mapper.java:145)_ > _at > org.apache.sqoop.mapreduce.AutoProgressMapper.run(AutoProgressMapper.java:64)_ > _at org.apache.hadoop.mapred.MapTask.runNewMapper(MapTask.java:787)_ > _at org.apache.hadoop.mapred.MapTask.run(MapTask.java:341)_ > _at org.apache.hadoop.mapred.YarnChild$2.run(YarnChild.java:164)_ > _at java.security.AccessController.doPrivileged(Native Method)_ > _at javax.security.auth.Subject.doAs(Subject.java:422)_ > _at > org.apache.hadoop.security.UserGroupInformation.doAs(UserGroupInformation.java:1657)_ > _at org.apache.hadoop.mapred.YarnChild.main(YarnChild.java:158)_ > *_Caused by: org.postgresql.util.PSQLException: Bad value for type double : > 100,000.00_* > _at org.postgresql.jdbc.PgResultSet.toDouble(PgResultSet.java:3104)_ > _at org.postgresql.jdbc.PgResultSet.getDouble(PgResultSet.java:2432)_ > _at > org.apache.sqoop.lib.JdbcWritableBridge.readDouble(JdbcWritableBridge.java:86)_ > _at > com.cloudera.sqoop.lib.JdbcWritableBridge.readDouble(JdbcWritableBridge.java:69)_ > _at tabletodrop.readFields(tabletodrop.java:106)_ > _at > org.apache.sqoop.mapreduce.db.DBRecordReader.nextKeyValue(DBRecordReader.java:244)_ > _... 12 more_ > However, other tables which doesn't have any field with datatype 'money' gets > imported without any issues. > Attached screenshot of the record in the postgresql table and the log > details for reference. -- This message was sent by Atlassian Jira (v8.3.4#803005)