Applied AI ML - Python & Agentic AI

JPMorgan Chase JPMorgan Chase · Banking · GLASGOW, LANARKSHIRE, United Kingdom · Commercial & Investment Bank

This role focuses on designing, building, and productionizing Generative AI and Agentic AI solutions, including LLM/SLM-powered applications like RAG, summarization, chat experiences, and tool-using agents. It requires strong software engineering skills in Java/Python, API development, MLOps, and cloud deployment on AWS, with an emphasis on secure coding, reliability, and automated evaluation.

What you'd actually do

  1. Design, develop, and deploy GenAI and Agentic AI solutions that improve automation, decision-making, and user experience across business workflows.
  2. Build LLM/SLM-powered applications including RAG-based systems, summarization/extraction pipelines, chat/coplay experiences, and tool-using agents.
  3. Engineer production-grade services using Java and/or Python (REST/gRPC APIs, microservices, libraries), following secure coding and reliability best practices.
  4. Develop prompt strategies and prompt engineering assets (templates, routing, guardrails), and implement automated evaluation to improve quality over time.
  5. Build and maintain data pipelines and processing workflows required for ML/GenAI use cases using cloud services.

Skills

Required

  • Undergrad or Master’s degree (or equivalent practical experience) in Computer Science, Data Science, Machine Learning, or related field.
  • Hands-on experience building applied AI/ML or GenAI solutions (e.g., RAG, classification, extraction, ranking, summarization, copilots).
  • Familiarity with MCP (Model Context Protocol), Agent Skills and architectures that connect models to tools/data through standardized interfaces.
  • Familiarity with LLM application patterns: embeddings/vector search, prompt orchestration, tool calling/function calling, safety/guardrails, evaluation.
  • Strong software engineering experience delivering production systems; ability to design maintainable architectures and write clean, testable code.
  • Proficiency in Java and/or Python and experience building APIs/services and integrating with data sources and downstream systems.
  • Experience deploying solutions on AWS and cloud-native environments; understanding of security fundamentals and operational excellence.
  • Experience with modern engineering practices: CI/CD, code reviews, unit testing (e.g., pytest/JUnit), and deployment automation.
  • Experience with containers and orchestration (e.g., Docker, Kubernetes/EKS) and production monitoring practices.

Nice to have

  • Experience building agentic AI systems (multi-step workflows, tool routing, planning, memory patterns, supervision/fallback strategies).
  • Experience with AWS Bedrock and/or SageMaker (or equivalent managed ML/GenAI platforms) and deployment patterns for scalable inference.
  • Experience with evaluation frameworks and approaches (golden datasets, LLM-as-judge, human-in-the-loop review, red teaming).
  • Experience fine-tuning models (e.g., LoRA/QLoRA/DoRA) and/or working with SLMs, embeddings, and retrieval systems.
  • Experience with developer productivity tooling such as GitHub Copilot and Claude Code, paired with strong SDLC controls.
  • Knowledge of the financial services industry and operating in regulated environments (auditability, controls, data handling).
  • Exposure to distributed compute/training concepts (e.g., DDP, sharding) and performance/cost optimization.

What the JD emphasized

  • Generative AI
  • Agentic AI
  • LLM/SLM
  • RAG
  • tool-using agents
  • evaluation
  • MLOps
  • backend/service engineering
  • Java
  • Python
  • APIs/microservices
  • testing
  • CI/CD
  • observability
  • reliability
  • AWS
  • cloud-native platforms
  • secure coding
  • reliability best practices
  • automated evaluation
  • data pipelines
  • processing workflows
  • experimentation
  • versioning
  • deployment
  • monitoring
  • maintenance
  • prompt engineering
  • prompt strategies
  • prompt routing
  • guardrails
  • performance benchmarking
  • latency/cost
  • logging/metrics/tracing
  • cross-functional stakeholders
  • technical and non-technical audiences
  • ambiguous environments
  • multiple stakeholders
  • MCP (Model Context Protocol)
  • Agent Skills
  • structured interfaces
  • embeddings/vector search
  • tool calling/function calling
  • safety/guardrails
  • maintainable architectures
  • clean, testable code
  • data sources
  • downstream systems
  • security fundamentals
  • operational excellence
  • code reviews
  • unit testing
  • deployment automation
  • containers
  • orchestration
  • Docker
  • Kubernetes/EKS
  • production monitoring practices
  • agentic AI systems
  • multi-step workflows
  • tool routing
  • planning
  • memory patterns
  • supervision/fallback strategies
  • AWS Bedrock
  • SageMaker
  • managed ML/GenAI platforms
  • scalable inference
  • evaluation frameworks
  • golden datasets
  • LLM-as-judge
  • human-in-the-loop review
  • red teaming
  • fine-tuning models
  • LoRA/QLoRA/DoRA
  • SLMs
  • embeddings
  • retrieval systems
  • developer productivity tooling
  • GitHub Copilot
  • Claude Code
  • SDLC controls
  • financial services industry
  • regulated environments
  • auditability
  • controls
  • data handling
  • distributed compute/training
  • DDP
  • sharding
  • performance/cost optimization

Other signals

  • Generative AI
  • Agentic AI
  • LLM/SLM
  • RAG
  • tool-using agents
  • evaluation
  • MLOps
  • backend/service engineering
  • AWS
  • cloud-native