My Pipeline Plan

Ever since I first started working in data visualisation I had been curious as to where data "came from ". Obviously, it comes from an Excel or a database. But how does it get into said place? What is the flow of data from a to be from the collection point to the dashboard? What are the pipes that carry it? In todays task, we planned out how we were going to move data from the mailchimp source systems all the way to a Tableau dashboard.

Step 1:

Extract via API callCall the Mailchimp API using a Python script. We extracted the data using a Python script with our secrets safely hidden away in a .env file. This gave us a series of JSON files, which we stored in a local folder.If we have time, we can also include error handeling and logging ie if the API call sucseece give one message; if it fails, check against a list of pre approved fail messages.Step 2: Set up our AWS ecosystem.We need to set up a couple of things inside our AWS system. They are

  • Role
  • User
  • Policy
  • Bucket

They will mesh together in order to provide the nessecery permissions in order for us to push from our local file to our S3 bucket.

Step 2: Load to AWS S3 Bucket

We then use a second Python script to load from our temporary file location to an S3 bucket. Again, we stash all of our credentials in an .env file to keep them out of the way. The reason why we want to go for an S3 bucket is that it is a very cheap and very flexible way in which nearly any kind of data can be stored.

Step 3: Set up Snowflake integration

Once we have our data in S3, we need to link it to our Snowflake instance, where we will be able to parse our data and build out our models. If you want to understand the process in more detail, I have already written a blog about it.

Snowflake intergrations. - The Data School
A classic data engineering task is to connect different systems together, in this blog we will run through the process of setting up the intergration of Snowflake with an S3 bucket. This will enable us to populate our snowflake instance with the data which we have in an S3 bucket

Step 4 Snowpipe

Once we have set up out S3 --> Snowflake intergration we can use a snowpipe to automate the procees. We tell it where to look ( our S3 bucket ) and it will check when ever something appears in there. Once it sees something new, it will copy it into our new table.

Author:
Olivier Newman
Powered by The Information Lab
1st Floor, 25 Watling Street, London, EC4M 9BR
Subscribe
to our Newsletter
Get the lastest news about The Data School and application tips
Subscribe now
© 2025 The Information Lab