Research Scientist, Frontier Red Team (autonomy)

Anthropic Anthropic · AI Frontier · AI Policy & Societal Impacts

Research Scientist role focused on developing and productionizing advanced autonomy evaluations for AI Safety Level (ASL) determination of models. This involves risk and capability modeling, designing, implementing, and running large-scale experiments to evaluate autonomous capabilities and forecast future capabilities, with potential for people management.

What you'd actually do

  1. Lead the end-to-end development of autonomy evals and research. This starts with risk and capability modeling, and includes designing, implementing, and regularly running these evals.
  2. Quickly iterate on experiments to evaluate autonomous capabilities and forecast future capabilities.
  3. Provide technical leadership to Research Engineers to scope + build scalable and secure infrastructure to quickly run large-scale experiments.
  4. Communicate the outcomes of the evaluations to relevant Anthropic teams, as well as policy stakeholders and research collaborators, where relevant.
  5. Collaborate with other projects on the Frontier Red Team, Alignment, and beyond to improve infrastructure and design safety techniques for autonomous capabilities.

Skills

Required

  • ML background
  • experience leading experimental research on LLMs/multimodal models and/or agents
  • strong Python-based engineering skills
  • design and run experiments
  • experience training, working with, and prompting models

Nice to have

  • driven to find solutions to ambiguously scoped problems
  • thrive in a collaborative environment
  • people management

What the JD emphasized

  • autonomy evals
  • agentic
  • autonomy evals
  • agentic
  • autonomy evals
  • agentic

Other signals

  • Developing and implementing advanced autonomy evals
  • Studying increasingly capable and agentic models
  • Forecasting future capabilities