Researcher, Safety Oversight

OpenAI OpenAI · AI Frontier · San Francisco, CA · Safety Systems

Researcher focused on AI safety, specifically developing methods for oversight of frontier AI models, identifying and mitigating misuse and misalignment, and improving models' reasoning about human values. The role involves developing AI monitor models, designing red-teaming pipelines, and collaborating with cross-functional teams.

What you'd actually do

  1. Develop and refine AI monitor models to detect and mitigate known and emerging patterns of misuse and misalignment.
  2. Set research directions and strategies to make our AI systems safer, more aligned, and more robust.
  3. Evaluate and design effective red-teaming pipelines to examine the end-to-end robustness of our safety systems, and identify areas for future improvement.
  4. Conduct research to improve models’ ability to reason about questions of human values, and apply these improved models to practical safety challenges.
  5. Coordinate and collaborate with cross-functional teams, including T&S, legal, policy and other research teams, to ensure that our products meet the highest safety standards.

Skills

Required

  • Python
  • AI safety research
  • RLHF
  • human-AI collaboration
  • fairness & biases
  • research engineering experience

Nice to have

  • Ph.D. or other degree in computer science, machine learning, or a related field
  • experience in safety research
  • experience in AI misuse and misalignment

What the JD emphasized

  • AI safety
  • AI misuse
  • AI misalignment
  • AI oversight
  • red-teaming pipelines
  • human values
  • safety standards

Other signals

  • AI safety
  • AI alignment
  • AI misuse
  • AI oversight
  • AI misuse and misalignment