Research Engineer, Reward Models Platform

Anthropic Anthropic · AI Frontier · United States · Remote · AI Research & Engineering

Research Engineer focused on building platforms and infrastructure to automate and accelerate the reward model development and evaluation workflows for ML researchers at Anthropic. The role involves creating tools for rubric development, human feedback analysis, reward robustness evaluation, and detecting reward hacks, with the goal of enabling rapid iteration and improving reward signal quality for training AI models.

What you'd actually do

  1. Design and build infrastructure that enables researchers to rapidly iterate on reward signals, including tools for rubric development, human feedback data analysis, and reward robustness evaluation
  2. Develop systems for automated quality assessment of rewards, including detection of reward hacks and other pathologies
  3. Create tooling that allows researchers to easily compare different reward methodologies (preference models, rubrics, programmatic rewards) and understand their effects
  4. Build pipelines and workflows that reduce toil in reward development, from dataset preparation to evaluation to deployment
  5. Implement monitoring and observability systems to track reward signal quality and surface issues during training runs

Skills

Required

  • Python
  • ML workflows
  • data pipelines
  • infrastructure development
  • tooling development
  • platform development
  • collaboration with researchers
  • ambiguous problem scoping
  • rapid iteration

Nice to have

  • ML research experience
  • internal tooling and platforms for ML researchers
  • data quality assessment
  • pipeline optimization
  • experiment tracking
  • evaluation frameworks
  • MLOps tooling
  • large-scale data processing
  • Kubernetes
  • distributed systems
  • cloud infrastructure
  • reinforcement learning
  • fine-tuning workflows

What the JD emphasized

  • strong engineering fundamentals
  • research experience
  • ship quickly
  • Python skills
  • ML workflows and data pipelines
  • building related infrastructure/tooling/platforms
  • working across the stack
  • results-oriented
  • impact

Other signals

  • building platforms for ML researchers
  • automating ML workflows
  • reward models
  • fine-tuning