Cross-functional Prompt Engineer

Anthropic Anthropic · AI Frontier · AI Research & Engineering

This role focuses on shaping and owning the behavior of Claude, Anthropic's AI model, across all products. It involves authoring system prompts, developing meta-prompts for research pipelines, leading incident response for behavioral issues, and scaling best practices. The role requires strong prompting skills, technical foundations, excellent judgment, and collaboration across research, product, and safety teams. It sits at the intersection of research and product, aiming to ensure AI systems are safe, beneficial, and aligned with human values at scale.

What you'd actually do

  1. Author and maintain behavior system prompts for each new Claude model release, ensuring consistent and aligned behaviors across products
  2. Deliver meta-prompts for critical research synthetic data pipelines, enabling our alignment and training efforts
  3. Review production prompt changes from product teams and serve as a resource for particularly challenging prompting problems involving alignment and reputational risks
  4. Identify, triage, and prioritize behavioral issues across Claude products, leading incident response for behavioral and policy concerns
  5. Develop behavioral evaluations in collaboration with product teams and alignment research to measure and track Claude's behaviors

Skills

Required

  • Prompt engineering experience with large language models
  • Writing and evaluating complex multi-step prompts
  • Deep knowledge of Claude's behaviors, capabilities, and limitations
  • Intuition for promptable issues vs. model-layer changes
  • Python
  • Creating behavioral evaluations from scratch
  • Judgment about model behaviors
  • Technical understanding (agent scaffold architectures, model training processes)
  • Collaboration across research, product, and safety teams
  • Core product management skills (prioritization, requirements gathering, stakeholder management, translating user feedback)
  • Independent execution and responsiveness
  • Understanding of AI safety and ethical implications

Nice to have

  • Background in philosophy, ethics, or psychology
  • Experience with RLHF, constitutional AI, or other alignment techniques
  • Track record of writing specifications or guidelines
  • Experience responding to safety incidents or behavioral issues in production AI systems
  • Formal training in ethics or moral philosophy
  • Published work or demonstrated expertise in AI safety or alignment
  • Experience building and maintaining evaluation frameworks for language models
  • Background in data science (data quality, verification)

What the JD emphasized

  • critical system prompts
  • critical research synthetic data pipelines
  • challenging prompting problems
  • behavioral issues
  • behavioral evaluations
  • prompt changes
  • model-specific prompt guides
  • prompt infrastructure improvements
  • prompting best practices
  • extensive prompt engineering experience
  • writing and evaluating complex multi-step prompts
  • deep knowledge of Claude's behaviors, capabilities, and limitations
  • strong intuition for what issues are promptable versus requiring model-layer changes
  • write Python
  • create behavioral evaluations from scratch
  • strong technical understanding
  • agent scaffold architectures
  • model training processes
  • core product management skills
  • independent drive changes through production systems
  • strong execution and responsiveness
  • AI safety
  • model welfare
  • ethical implications of model behaviors

Other signals

  • own and shape Claude's behaviors across all of our products
  • authoring critical system prompts for new models and products
  • delivering complex meta-prompts that drive our research pipelines
  • lead the response to behavioral issues
  • bring rigor to our production prompts and scaling best practices