Research Engineer, Societal Impacts

Anthropic Anthropic · AI Frontier · AI Policy & Societal Impacts

Research Engineer focused on building infrastructure for studying the societal impacts of AI systems, including economic, wellbeing, and educational effects, as well as socio-technical alignment and novel capability evaluation. The role involves designing and implementing scalable technical infrastructure for experiments, data pipelines, and monitoring systems, working closely with researchers and cross-functional partners to generate insights and advance AI safety.

What you'd actually do

  1. Design and implement scalable technical infrastructure that enables researchers to efficiently run experiments and evaluate AI systems
  2. Architect systems that can handle uncertain and changing requirements while maintaining high standards of reliability
  3. Lead technical design discussions to ensure our infrastructure can support both current needs and future research directions
  4. Partner closely with researchers, data scientists, policy experts, and other cross-functional partners to advance Anthropic’s safety mission
  5. Interface with and improve our internal technical infrastructure and tools
  6. Generate net-new insights about the potential societal impact of systems being developed by Anthropic

Skills

Required

  • Python
  • design and implement scalable technical infrastructure
  • architect systems
  • working with Research Scientists on ambiguous AI projects
  • building and maintaining production-grade internal tools or research infrastructure
  • writing clean, well-documented code
  • making technical decisions with incomplete information
  • getting up-to-speed quickly on unfamiliar codebases
  • using technical infrastructure to interface effectively with machine learning models
  • deriving insights from imperfect data streams

Nice to have

  • Maintaining large, foundational infrastructure
  • Building simple interfaces that allow non-technical collaborators to evaluate AI systems
  • Working with and prioritizing requests from a wide variety of stakeholders
  • distributed systems
  • design for scale and reliability
  • Scaling and optimizing the performance of tools

What the JD emphasized

  • track record of running & designing experiments relating to machine learning systems
  • building data processing pipelines
  • architecting & implementing high-quality internal infrastructure
  • navigating the ambiguity inherent to novel empirical research
  • eagerness to develop their own research & technical skills
  • building and maintaining production-grade internal tools or research infrastructure
  • using technical infrastructure to interface effectively with machine learning models
  • deriving insights from imperfect data streams

Other signals

  • design and build critical infrastructure that enables and accelerates foundational research into how our AI systems impact people and society
  • rigorous empirical methods with creative technical approaches
  • design and implement scalable technical infrastructure that enables researchers to efficiently run experiments and evaluate AI systems
  • architect systems that can handle uncertain and changing requirements while maintaining high standards of reliability
  • partner closely with researchers, data scientists, policy experts, and other cross-functional partners to advance Anthropic’s safety mission
  • generate net-new insights about the potential societal impact of systems being developed by Anthropic
  • track record of running & designing experiments relating to machine learning systems
  • building data processing pipelines
  • architecting & implementing high-quality internal infrastructure
  • navigating the ambiguity inherent to novel empirical research
  • eagerness to develop their own research & technical skills
  • mixture of running experiments, developing new tools & evaluation suites
  • working cross-functionally across multiple research and product teams
  • striving for beneficial & safe uses for AI
  • building and maintaining production-grade internal tools or research infrastructure
  • using technical infrastructure to interface effectively with machine learning models
  • deriving insights from imperfect data streams
  • building simple interfaces that allow non-technical collaborators to evaluate AI systems
  • design and implement scalable infrastructure for running large-scale experiments on how people interact with our AI systems
  • build robust monitoring systems that help us detect and understand potential misuse or unexpected behaviors
  • create internal tools that help researchers, policy experts, and product teams quickly analyze dynamically changing AI system characteristics