From Petl
The connector can be used to create ETL applications and pipelines for CSV data in Python using Petl.
Install Required Modules
Install the Petl modules using the pip utility.pip install petl
Connecting
After you import the modules, including the CData Python Connector for Amazon Redshift, you can use the connector's connect function to create a connection using a valid Amazon Redshift connection string. If you prefer not to use a direct connection, you can use a SQLAlchemy engine.import petl as etl import cdata.redshift as mod cnxn = mod.connect("User=admin;Password=admin;Database=dev;Server=examplecluster.my.us-west-2.redshift.amazonaws.com;Port=5439;")
Extract, Transform, and Load the Amazon Redshift Data
Create a SQL query string and store the query results in a DataFrame.sql = "SELECT ShipName, ShipCity FROM \"sales_db\".\"public\".Orders " table1 = etl.fromdb(cnxn,sql)
Loading Data
With the query results stored in a DataFrame, you can load your data into any supported Petl destination. The following example loads the data into a CSV file.etl.tocsv(table1,'output.csv')
Modifying Data
Insert new rows into Amazon Redshift tables using Petl's appenddb function.table1 = [['ShipName','ShipCity'],['Raleigh','New York']] etl.appenddb(table1,cnxn,'\"sales_db\".\"public\".Orders')