Research Engineer, Model Evaluations

Anthropic Anthropic · AI Frontier · AI Research & Engineering

Research Engineer focused on designing and implementing Anthropic's model evaluation platform, shaping how models are understood, measured, and improved. This role involves leading the architecture of scalable evaluation infrastructure, implementing high-throughput pipelines for production training, analyzing results to guide model development, and collaborating with research and training teams. The goal is to ensure models meet high standards for capabilities and safety before deployment, influencing training decisions and the overall model roadmap.

What you'd actually do

  1. Design novel evaluation methodologies to assess model capabilities across diverse domains including reasoning, safety, helpfulness, and harmlessness
  2. Lead the design and architecture of Anthropic's evaluation platform, ensuring it scales with our rapidly evolving model capabilities and research needs
  3. Implement and maintain high-throughput evaluation pipelines that run during production training, providing real-time insights to guide training decisions
  4. Analyze evaluation results to identify patterns, failure modes, and opportunities for model improvement, translating complex findings into actionable insights
  5. Partner with research teams to develop domain-specific evaluations that probe for emerging capabilities and potential risks

Skills

Required

  • Python
  • distributed computing frameworks
  • systems engineering
  • experimental design
  • statistical analysis
  • technical leadership

Nice to have

  • reinforcement learning evaluation
  • multi-agent systems
  • psychometrics
  • experimental psychology
  • prompt engineering
  • red teaming methodologies

What the JD emphasized

  • critical system that shapes how we understand, measure, and improve our models' capabilities and safety
  • directly influences our training decisions and model development roadmap
  • technical leadership role
  • lead the design and architecture
  • high-throughput evaluation pipelines that run during production training
  • Analyze evaluation results to identify patterns, failure modes, and opportunities for model improvement
  • Partner with research teams to develop domain-specific evaluations that probe for emerging capabilities and potential risks
  • Build infrastructure to enable rapid iteration on evaluation design
  • Establish best practices and standards for evaluation development
  • Coordinate evaluation efforts during critical training runs
  • Have experience designing and implementing evaluation systems for machine learning models, particularly large language models
  • demonstrated technical leadership experience
  • skilled at both systems engineering and experimental design
  • translate between research needs and engineering constraints
  • results-oriented and thrive in fast-paced environments where priorities can shift based on research findings
  • Experience with evaluation during model training, particularly in production environments
  • safety evaluation frameworks and red teaming methodologies
  • managing evaluation infrastructure at scale

Other signals

  • design and implementation of Anthropic's evaluation platform
  • develop and implement model evaluations
  • influences our training decisions and model development roadmap
  • high-throughput evaluation pipelines that run during production training
  • Analyze evaluation results to identify patterns, failure modes, and opportunities for model improvement
  • Partner with research teams to develop domain-specific evaluations that probe for emerging capabilities and potential risks
  • Build infrastructure to enable rapid iteration on evaluation design
  • Establish best practices and standards for evaluation development
  • Coordinate evaluation efforts during critical training runs
  • Contribute to research publications