Abuse Investigator (ai Self-improvement Risk)

OpenAI OpenAI · AI Frontier · San Francisco, CA · Intelligence & Investigations

Investigate AI agentic and autonomous behavior risks, identify emerging patterns, and propose improvements to safeguards. Requires expertise in complex systems, multi-step behaviors, and ambiguous signals.

What you'd actually do

  1. Review leads, investigate model behavior, and identify cases where systems demonstrate agentic or autonomous patterns that that introduce safety risks
  2. Detect and analyze behaviors such as multi-step planning, capability chaining, tool use, persistence, and workaround behavior
  3. Develop signals and tracking strategies to help proactively identify emerging agentic risk patterns across our platform
  4. Identify gaps in existing safeguards, evaluations, or monitoring systems and propose improvements
  5. Communicate investigation findings clearly to technical, policy, and leadership stakeholders

Skills

Required

  • SQL
  • Python

Nice to have

  • investigating complex, adversarial, or emergent system behavior
  • AI safety
  • security
  • cyber
  • trust & safety environments
  • technical investigations
  • analyzing multi-step systems, automation, or agentic workflows
  • threat analysis
  • research in complex and ambiguous domains
  • identifying failure modes, unintended behaviors, or system-level risks
  • developing automated or scalable approaches to detection or investigation
  • presenting analytic work in technical, research, or policy settings

What the JD emphasized

  • deep domain-specific expertise in identifying, understanding, and mitigating risk from agentic systems, model autonomy, and AI self-improvement signals
  • experience investigating complex systems where behavior emerges across multiple steps, tools, or interactions
  • distinguish between normal task execution and concerning patterns such as persistence, workaround behavior, or capability expansion
  • proven ability to navigate ambiguous signals in a rapidly evolving and highly technical environment
  • reviewing complex or sensitive model behaviors and edge-case outputs, requiring strong judgment and resilience in high-pressure environments
  • deep expertise in investigating complex, adversarial, or emergent system behavior, ideally in AI safety, security, cyber, or trust & safety environments
  • experience analyzing multi-step systems, automation, or agentic workflows, and understanding how behaviors emerge across interactions
  • at least 6 years of experience conducting investigations, threat analysis, or research in complex and ambiguous domains
  • experience identifying failure modes, unintended behaviors, or system-level risks, particularly in AI or software systems
  • at least two years of experience helping to develop automated or scalable approaches to detection or investigation

Other signals

  • identifying and investigating misuses of our products
  • identifying and investigating cases where models exhibit autonomous or agentic behavior
  • detecting behaviors that are not explicitly intended, understood, or covered by existing safeguards
  • deep domain-specific expertise in identifying, understanding, and mitigating risk from agentic systems, model autonomy, and AI self-improvement signals
  • experience investigating complex systems where behavior emerges across multiple steps, tools, or interactions
  • distinguish between normal task execution and concerning patterns such as persistence, workaround behavior, or capability expansion
  • navigate ambiguous signals in a rapidly evolving and highly technical environment
  • reviewing complex or sensitive model behaviors and edge-case outputs
  • strong judgment and resilience in high-pressure environments