Anthropic AI Safety Fellow, UK

Anthropic Anthropic · AI Frontier · AI Research & Engineering

This is a research fellowship focused on AI safety, aiming to produce empirical research outputs like paper submissions. Fellows will use external infrastructure and work on projects aligned with Anthropic's research priorities, receiving mentorship and resources.

What you'd actually do

  1. Fellows will use external infrastructure (e.g. open-source models, public APIs) to work on an empirical project aligned with our research priorities, with the goal of producing a public output (e.g. a paper submission).
  2. Fellows will receive substantial support - including mentorship from Anthropic researchers, funding, compute resources, and access to a shared workspace - enabling them to develop the skills to contribute meaningfully to critical AI safety research.
  3. Direct mentorship from Anthropic researchers
  4. Connection to the broader AI safety research community

Skills

Required

  • strong technical background in computer science, mathematics, physics, or related fields
  • strong programming skills, particularly in Python
  • machine learning frameworks
  • comfortable programming in Python
  • work full-time on the fellowship for at least 2 months, and ideally 6 months
  • work authorisation (US, UK, or Canada)
  • work out of Berkeley or London (or remotely if in Canada)

Nice to have

  • Experience with empirical ML research projects
  • Experience working with Large Language Models
  • Experience in one of the research areas

What the JD emphasized

  • AI safety research
  • empirical project
  • paper submission

Other signals

  • AI safety research
  • empirical project
  • paper submission