This is currently a volunteer position. Contributors will gain hands-on experience building real-world AI systems for public health impact.
About the Role
This unpaid volunteer role oversees evaluation, safety, and responsible AI governance for LLM-powered healthcare chatbot systems.
Commitment
- 4–6 hours per week
- Remote
- Flexible schedule
Qualifications
- Background in data science, NLP, machine learning, or AI research
- Experience with LLM evaluation, RAG, and prompt engineering
- Understanding of model validation and bias mitigation
- Interest in healthcare ethics and responsible AI
Responsibilities
- Design evaluation framework for chatbot outputs
- Develop medical safety test cases
- Measure hallucination rates and factual accuracy
- Implement guardrails and bias monitoring
- Document governance practices