Generative AI Director

JPMorgan Chase JPMorgan Chase · Banking · LONDON, LONDON, United Kingdom · Commercial & Investment Bank

Lead the design and delivery of production-grade LLM systems for mission-critical financial products, focusing on scalable APIs, agentic workflows, and end-to-end ownership of LLM Suite products. This role bridges AI research with robust engineering for enterprise-wide AI adoption.

What you'd actually do

  1. You will architect scalable APIs and agentic workflows, enabling automation and efficiency across the firm.
  2. Architect and deliver production LLM-based systems for text, image, speech, and video applications.
  3. Own end-to-end delivery, performance, and continuous improvement of LLM Suite products.
  4. Working closely with ML Engineering, Product Management, and Cloud Engineering, you will ensure our AI solutions are reliable, secure, and built for real business impact.
  5. Bridge advanced AI research with robust engineering to build innovative, production-ready solutions.

Skills

Required

  • PhD or equivalent experience in Computer Science, Mathematics, Statistics, or a related quantitative discipline
  • Extensive hands-on experience in ML engineering
  • Proven track record of shipping production AI systems
  • Deep expertise in NLP, Computer Vision, and/or Multimodal LLM algorithms
  • Strong foundation in statistics, optimization, and ML theory
  • Practical experience implementing distributed, multi-threaded, and scalable applications using frameworks such as Ray, Horovod, or DeepSpeed
  • Communicate complex technical concepts effectively
  • Build trust with stakeholders at all levels

Nice to have

  • Design and deploy production ML pipelines using DAG frameworks, including custom operator development and pipeline optimization
  • Architect and implement high-throughput, low-latency microservices with gRPC, REST, and GraphQL, including protocol buffer schema design, streaming endpoints, and load balancing
  • Hands-on experience with parameter-efficient fine-tuning (LoRA, QLoRA, IA3), model quantization (INT8, FP16, GPTQ), and quantization-aware training for LLMs at scale
  • Deep knowledge of distributed training strategies, memory optimization, and inference acceleration for large-scale multimodal models
  • Orchestrate advanced agentic workflows, including multi-agent coordination, stateful task management, and integration with enterprise event-driven architectures

What the JD emphasized

  • production-grade LLM systems
  • architect scalable APIs and agentic workflows
  • Architect and deliver production LLM-based systems
  • Own end-to-end delivery, performance, and continuous improvement of LLM Suite products
  • extensive hands-on experience in ML engineering, with a proven track record of shipping production AI systems
  • Apply practical experience implementing distributed, multi-threaded, and scalable applications
  • Architect and implement high-throughput, low-latency microservices
  • Orchestrate advanced agentic workflows

Other signals

  • architect scalable APIs and agentic workflows
  • architect and deliver production LLM-based systems
  • own end-to-end delivery, performance, and continuous improvement of LLM Suite products
  • bridge advanced AI research with robust engineering
  • demonstrate extensive hands-on experience in ML engineering, with a proven track record of shipping production AI systems