On 01.11.2018 18:27, Ravi Krishna wrote:
I have a project to develop a script/tool to copy data from DB2 to PG. The
approach I am thinking is
1. Export data from db2 in a text file, with, say pipe as delimiter.
2. Load the data from the text file to PG using COPY command.
In order to make it faster I can parallelize export and load with upto X number
of tables concurrently.
Is there a tool in which I can avoid first exporting and then loading. I think
one way I can do it is by
using pipes by which I can export from db2 to a pipe and simultaneously load it
to PG using COPY.
Any other tool for this job?
thanks.
Haven't tried it myself, but you may be able to connect the DB2 database
to your PostgreSQL cluster using this FDW module:
https://github.com/wolfgangbrandl/db2_fdw
Then you could just use INSERT INTO ... SELECTÂ statements to do the ETL
process with the necessary type conversions and whatnot.
Looks like db2_fdw is DB2 LUW only though, so you might be out of luck
if your DB2 is on IBM i (or z ;-)
Kind regards
Florian