Researcher, Misalignment Research

OpenAI OpenAI · AI Frontier · San Francisco, CA · Safety Systems

Researcher focused on identifying, quantifying, and understanding future AGI misalignment risks. The role involves designing worst-case demonstrations, developing adversarial and system-level evaluations, creating automated red-teaming infrastructure, researching alignment technique failure modes, and publishing findings to influence safety strategy and product safeguards.

What you'd actually do

  1. Design and implement worst-case demonstrations that make AGI alignment risks concrete for stakeholders, focused on high stakes use cases described above.
  2. Develop adversarial and system-level evaluations grounded in those demonstrations, driving adoption across OpenAI.
  3. Create automated tools and infrastructure to scale automated red-teaming and stress testing.
  4. Conduct research on failure modes of alignment techniques and propose improvements.
  5. Publish influential internal or external papers that shift safety strategy or industry practice. We aim to concretely reduce existential AI risk.

Skills

Required

  • AI red-teaming
  • security research
  • adversarial ML
  • modern ML / AI techniques
  • large-scale codebases
  • evaluation infrastructure
  • clear communication with technical and non-technical audiences
  • collaboration
  • cross-functional project driving

Nice to have

  • Ph.D., master’s degree, or equivalent experience in computer science, machine learning, security, or a related discipline

What the JD emphasized

  • 4+ years of experience in AI red-teaming, security research, adversarial ML, or related safety fields
  • strong research track record—publications, open‑source projects, or high‑impact internal work

Other signals

  • AGI misalignment risks
  • adversarial evaluations
  • system-level stress testing
  • alignment techniques failure modes
  • existential AI risk