Data Automation Intern
Location: Fully Remote
Duration: 3 months (extendable)
About Us:
We are a non-profit organization using data to improve operations. This internship is a real-world, hands-on opportunity to design and run automated data pipelines and work directly with live data sources.
Role Overview:
As a Data Automation Intern, you will independently design, build, and maintain automated data connections between platforms (Kobo, Google Sheets) and BigQuery, exploring free/open-source tools to make processes smoother.
Responsibilities:
- Build and run automated data pipelines to BigQuery.
- Design ETL processes to collect, clean, and transform data.
- Debug and maintain existing pipelines to ensure reliable data flow.
- Document workflows, setups, and troubleshooting steps.
- Test pipelines using sample or low-usage datasets.
- Share weekly progress updates and actively participate in meetings.
- Explore new tools and automation techniques to improve efficiency.
Skills & Tools:
- Python: Practical scripting skills for CRUD operations, data transformations, and error handling.
- Comfortable exploring new tools (free/open-source or NPO-friendly).
- Familiarity with GCP, BigQuery, Google Sheets, APIs, Retool is a plus.
- Strong problem-solving and debugging abilities.
- Self-motivated and able to take ownership of tasks independently.
Deliverables / Outcome Expectations:
- Fully functional automated data pipelines built independently.
- Validated and tested data flows.
- Clear documentation of setups and troubleshooting processes.
- Weekly progress updates and completion of assigned tasks.
Why Join:
- Work independently on real-world automation projects.
- Hands-on learning in data pipelines, connectivity, and Python scripting.
- Flexible, fully remote internship with the chance to extend based on performance.
- Contribute directly to critical operational data systems for an NPO.