Click here to Skip to main content
15,943,365 members
Please Sign up or sign in to vote.
0.00/5 (No votes)
See more:
Problem : I have a source schema (a bunch of tables) and a destination schema (a fixed number of tables with predefined columns) in postgres. I want to design a mapper-insert function which would map the source column(s) to a particular column of a table in destination schema.

What I have tried:

Idea : I thought of constructing a JSON file which will be of {Destination_column : Source_column(s)} kind of construct. In case of multiple source column I thought of passing the as list in the value of dictionary. After that I plan to create a mapper function to define the operations on the source column and insert function to insert the value after performing the operation.

Question : Is this the correct pythonic way to solve this problem? Do we have any python libraries (petl or bonobo) which does source-destination mapping? Thoughts?
Updated 1-Feb-20 22:24pm

1 solution

Take a look at Orange: etl-tools-for-programmers~orange[^]
Share this answer

This content, along with any associated source code and files, is licensed under The Code Project Open License (CPOL)

CodeProject, 20 Bay Street, 11th Floor Toronto, Ontario, Canada M5J 2N8 +1 (416) 849-8900