
Job Overview
Location
London, UK; Ontario, CAN; Remote-Friendly, United States; San Francisco, CA
Job Type
Full-time
Category
Data Science
Date Posted
May 6, 2026
Full Job Description
📋 Description
- • The Anthropic Fellows Program — AI Safety is a 4-month full-time research fellowship designed to foster technical talent in AI safety and alignment, offering mentorship, funding, and the opportunity to produce public research outputs such as paper submissions.
- • Fellows will work on empirical projects aligned with Anthropic’s research priorities, including scalable oversight, adversarial robustness, model organisms, mechanistic interpretability, and AI welfare, under direct mentorship from leading researchers in the field.
- • The program is hosted by Anthropic, a public benefit corporation committed to building reliable, interpretable, and steerable AI systems, with shared workspaces in Berkeley, California, and London, UK, and remote options available for eligible candidates in the US, UK, or Canada.
- • Participants will gain hands-on experience in cutting-edge AI safety research, develop technical and communication skills, collaborate with a diverse team of researchers and engineers, and may be considered for full-time roles at Anthropic based on performance.
🎯 Requirements
- • Fluent in Python programming
- • Available to work full-time on the Fellows program for 4 months
- • Have work authorization in the US, UK, or Canada and be located in that country during the program
🏖️ Benefits
- • Weekly stipend of 3,850 USD / 2,310 GBP / 4,300 CAD (varies by country)
- • Funding for compute (~$15k/month) and other research expenses
- • Access to shared workspaces in Berkeley or London, or remote participation options
- • Direct mentorship from Anthropic researchers and connection to the broader AI safety research community
Skills & Technologies
About Anthropic, PBC
Anthropic is a public benefit corporation founded in 2021 by former OpenAI researchers to develop large-scale AI systems that are safe, interpretable and aligned with human values. The company produces Claude, a family of conversational and reasoning models based on constitutional AI and reinforcement learning from human feedback. Headquartered in San Francisco, Anthropic combines frontier research with applied engineering, publishing scholarly papers on alignment, interpretability and robustness while offering API access and commercial products built on its models.
Subscribe to the weekly newsletter for similar remote roles and curated hiring updates.
Newsletter
Weekly remote jobs and featured talent.
No spam. Only curated remote roles and product updates. You can unsubscribe anytime.
Similar Opportunities

FundraiseUp Inc.
2 months ago

Hangar Aviation Technologies, Inc.
2 months ago

Anyone AI Inc.
16 days ago
