Senior Applied Scientist – AI Red Teaming & Model Risk

Uber Uber · Consumer · New York, NY +2 · Data Science

This role focuses on AI Red Teaming and Model Risk for LLMs and agentic AI systems. The scientist will design and execute experiments to uncover unsafe or harmful behaviors, develop evaluation frameworks, and define risk metrics beyond standard accuracy. They will analyze agent workflows and collaborate with security and platform teams to implement guardrails and mitigations. The role requires experience with LLMs, adversarial evaluation, and analyzing complex model behavior.

What you'd actually do

  1. Design and execute AI red-teaming experiments against LLMs and AI agents to identify: prompt injection (direct & indirect), jailbreaking and policy bypass, model and tool poisoning, context and memory poisoning, behavioral drift and unsafe autonomy
  2. Develop adversarial datasets, probes, and test harnesses to systematically evaluate model and agent behavior under attack.
  3. Define and track AI risk metrics beyond accuracy (e.g., failure rates, drift indicators, unsafe action likelihood, confidence miscalibration).
  4. Analyze agent workflows and decision traces to understand how failures emerge across multi-step reasoning and tool use.
  5. Collaborate with security engineers and AI platform teams to translate findings into guardrails, mitigations, and design improvements.

Skills

Required

  • 5+ years of experience as a Data Scientist, Applied Scientist, or ML Scientist
  • Hands-on experience working with LLMs or generative AI systems
  • Direct experience with AI red teaming, model safety, or adversarial evaluation
  • Direct experience with prompt injection, jailbreaks, and LLM failure modes
  • Strong background in experimental design, evaluation, and statistical analysis
  • Experience analyzing complex model behavior and failure cases beyond standard metrics
  • Proficiency in Python and common DS/ML tooling

Nice to have

  • Experience evaluating agentic systems, including tool use, memory, or multi-step workflows
  • Knowledge of GenAI architectures (transformers, embeddings, RAG, agent frameworks)
  • Experience building custom evaluation datasets or simulation environments
  • Background or strong interest in security, privacy, or trust & safety
  • Familiarity with AI evaluation tools (e.g., custom judges, LLM-as-judge, simulation frameworks)

What the JD emphasized

  • AI red teaming
  • model safety
  • adversarial evaluation
  • prompt injection
  • jailbreaks
  • LLM failure modes
  • evaluating agentic systems
  • tool use
  • multi-step workflows

Other signals

  • AI Red Teaming
  • Adversarial Evaluation
  • Model Risk
  • LLM Failure Analysis
  • Agentic AI Safety