AI Agents Applied Research/engineering Lead - Vice President

JPMorgan Chase JPMorgan Chase · Banking · Brooklyn, NY +1 · Consumer & Community Banking

Lead the research and end-to-end lifecycle of LLM-based agents for financial tasks, focusing on multi-step planning, tool use, safety, and production deployment under strict regulatory constraints. This role involves fine-tuning, optimizing LLMs, applying reinforcement learning, scaling systems, and implementing robust safety and compliance controls.

What you'd actually do

  1. Lead research and deployment of agentic AI systems with multi-step workflows, tool calling, and multi-agent orchestration.
  2. Fine-tune and optimize LLMs using parameter-efficient fine-tuning (PEFT), distillation, and quantization to meet production constraints such as latency, memory, and cost.
  3. Apply reinforcement learning and preference optimization to improve personalization and dialogue policies.
  4. Scale LLM systems through caching, batching, prompt governance, and evaluation frameworks.
  5. Implement privacy, safety, and security controls including PCI compliance, jailbreak resistance, and auditability.

Skills

Required

  • Ph.D. with 1+ years or M.S. with 3+ years building and deploying AI systems in production
  • Applied GenAI experience with LLMs including fine-tuning, prompt engineering, and RAG.
  • Experience scaling LLM systems with caching, batching, governance, and evaluation.
  • Strong foundation in ML, deep learning, statistical modeling, and experimental design.
  • Experience in Information Retrieval (indexing, ranking, retrieval) and/or recommendation systems.
  • Proficiency in Python and ML frameworks (PyTorch/TensorFlow, Hugging Face, scikit-learn)
  • Demonstrated ability to set a technical research agenda and drive it from concept through production deployment.
  • Experience presenting research findings and technical strategy to senior leadership and non-technical stakeholders.

Nice to have

  • 5+ years developing conversational AI systems, virtual assistants or LLM-based systems in production.
  • Experience with multi-agent orchestration, supervisor agents, and specialized toolkits.
  • Expertise in agent governance, red-teaming, adversarial testing, and safety evaluation.
  • Experience with reinforcement learning, bandit algorithms, and preference-based optimization (DPO, IPO), with practical exposure to data collection, labeling, and evaluation pipelines.
  • MLOps/LLMOps experience with CI/CD, monitoring, versioning, A/B testing, and rollbacks.
  • Track record of data-driven product development and experimentation.
  • Publications in top-tier AI/ML venues and/or open-source contributions

What the JD emphasized

  • production constraints
  • latency
  • accuracy
  • compliance constraints
  • auditable
  • explainable
  • safe
  • highly regulated
  • high-stakes domain
  • PCI compliance
  • jailbreak resistance
  • auditability

Other signals

  • LLM-based agents
  • multi-step planning
  • tool use
  • safety
  • production systems
  • real-world latency, accuracy, and compliance constraints
  • auditable, explainable, and safe in a highly regulated, high-stakes domain