Protection Scientist Engineer, Intelligence and Investigations

OpenAI OpenAI · AI Frontier · San Francisco, CA · Intelligence & Investigations

This role focuses on designing and building systems to identify and enforce abuse on OpenAI's products, leveraging data science, ML, and investigation. It involves developing monitoring for new and existing products, prototyping detection and enforcement systems, and investigating critical escalations. The role requires experience with SQL, Python, data engineering, ML principles, and scaling processes with language models.

What you'd actually do

  1. Scope and implement abuse monitoring requirements for new product launches.
  2. Improve processes to sustain monitoring operations for existing products, including developing approaches to automate monitoring subtasks.
  3. Prototype and mature into production systems of detection, review, and enforcement of abuse for major harms.
  4. Work with Product, Policy, Ops, and Investigative teams to understand key risks and how to address them, and with Engineering teams to ensure we have sufficient data and scaled tooling.

Skills

Required

  • 4 years of experience doing technical analysis and detection
  • SQL
  • Python
  • experience in trust and safety
  • experience with basic data engineering
  • machine learning principles and execution

Nice to have

  • investigative mindset
  • Basic software development skills
  • worked closely with policy, enforcement, and engineering teams

What the JD emphasized

  • designing and building systems to proactively identify and enforce on abuse
  • prototyping and incubating systems of defense against our highest risk harms
  • expert understanding of our products and data
  • scaling and automating processes, especially with language models

Other signals

  • designing and building systems to proactively identify and enforce on abuse
  • prototyping and incubating systems of defense against our highest risk harms
  • automating monitoring subtasks
  • scaling and automating processes, especially with language models