Effective altruism
opportunities board
Work on the world's most pressing problems. Browse jobs, fellowships, internships, courses, and more at high-impact organisations.
Deadline
2026-04-30
Routes to impact
Direct high impact on an important cause
Skill-building & building career capital
Description
A paid, short-term contract for engineers and researchers to build and test LLM evaluation benchmarks, suited for those with strong technical and research experience in AI systems.
- Scope Lead or support end-to-end development of LLM benchmarks, including evaluation design, scoring methods, and testing pipelines
- Skills Work with LLM APIs, prompt engineering, statistical analysis, and reproducible research code
- Compensation $40–$150/hour depending on role and experience, with flexible part-time or near full-time options
- Timeline 4-month contract starting May–July 2026, with priority applications by April 30
This text was generated by AI. If you notice any inconsistencies, please let us know using this form
Related opportunities
Contributor
EquistampRemtoe
Today
Research Scientist
Truthful AIBerkeley, California (preferred) or remote (US or international)
6 days ago
AI Safety Research Scientist
Beneficial AI FoundationRemote from anywhere or Cambridge, Massachusetts
1 month ago
List of Internship & Residency Programs – ML and Software Engineering [Compiled by 80,000 Hours]
80,000 Hours (80K)Mixed Locations
8 months ago
Join 60k subscribers and sign up for the EA Newsletter, a monthly email with the latest ideas and opportunities