Generative AI - Vice President

JPMorgan Chase JPMorgan Chase · Banking · LONDON, LONDON, United Kingdom · Corporate Sector

This role focuses on architecting and scaling production-grade LLM systems, including APIs and agentic workflows, for enterprise-wide adoption in a financial technology setting. It emphasizes end-to-end delivery, performance, and continuous improvement, bridging AI research with robust engineering for real business impact.

What you'd actually do

  1. Architect and deliver production LLM based systems (text, image, speech, video) powering mission-critical LLM Suite products.
  2. Own end-to-end delivery, performance, and continuous improvement of individual LLM Suite products.
  3. Bridge advanced AI research with robust engineering to build innovative, production-ready solutions.
  4. Drive results with an entrepreneurial mindset in a fast-paced, high-impact environment.

Skills

Required

  • PhD or equivalent experience in Computer Science, Mathematics, Statistics, or a related quantitative discipline.
  • Extensive hands-on experience as an individual contributor in ML engineering
  • proven track record of shipping production AI systems
  • Deep expertise in NLP, Computer Vision, and/or Multimodal LLM algorithms
  • strong foundation in statistics, optimization, and ML theory
  • Practical experience implementing distributed, multi-threaded, and scalable applications using frameworks such as Ray, Horovod, DeepSpeed, etc.
  • Exceptional communication skills

Nice to have

  • Advanced proficiency in designing and deploying production ML pipelines using DAG frameworks, including custom operator development and pipeline optimization.
  • Expertise in architecting and implementing high-throughput, low-latency microservices with gRPC, REST, and GraphQL, including protocol buffer schema design, streaming endpoints, and load balancing.
  • Hands-on experience with parameter-efficient fine-tuning (LoRA, QLoRA, IA3), model quantization (INT8, FP16, GPTQ), and quantization-aware training for LLMs at scale.
  • Deep knowledge of distributed training strategies (data/model/pipe parallelism), memory optimization, and inference acceleration for large-scale multimodal models.
  • Experience with advanced agentic workflow orchestration, including multi-agent coordination, stateful task management, and integration with enterprise event-driven architectures.

What the JD emphasized

  • production-grade LLM systems
  • agentic workflows
  • production AI systems
  • production LLM based systems
  • production-ready solutions
  • high-throughput, low-latency microservices
  • advanced agentic workflow orchestration

Other signals

  • production-grade LLM systems
  • scale robust, reusable APIs and agentic workflows
  • deliver solutions that are measured, budgeted, and built for real business impact
  • architect and deliver production LLM based systems
  • Own end-to-end delivery, performance, and continuous improvement
  • Bridge advanced AI research with robust engineering
  • ship production AI systems
  • implementing distributed, multi-threaded, and scalable applications
  • architecting and implementing high-throughput, low-latency microservices
  • Deep knowledge of distributed training strategies
  • Experience with advanced agentic workflow orchestration