Modal
ScalingData AI · GPU cloud
Currently tracking 7 active AI roles, up 175% versus the prior 4 weeks. Primary focus: Serve · Engineering.
Hiring
7 / 7
Momentum (4w)
↑+7 +175%
11 opens last 4w · 4 prior 4w
Salary range
—
Tracked since
Sep '25
last role today
Hiring velocityscroll left for older weeks
Jobs (7)
| Title | Stage | AI score |
|---|---|---|
| Member of Technical Staff - ML Performance Seeking an ML Performance Engineer with 5+ years of experience to optimize ML systems for higher throughput and lower latency. The role involves working with inference engines like vLLM or TensorRT, understanding GPU architecture, and improving ML performance at scale. | Serve | 8 |
| Member of Technical Staff - Agent DX Research Research role focused on building an evaluation framework for AI coding agents to improve developer experience on the Modal platform. This involves defining quantitative objectives, measuring performance, and translating insights into product improvements, while staying updated on agent advancements and customer use cases. | AgentEval Gate | 8 |
| Forward Deployed Engineer - ML Forward Deployed ML Engineer to partner with leading AI companies and foundation model labs to help them achieve state-of-the-art performance on demanding workloads like LLM serving, model training (SFT, RLHF), and audio pipelines. This role involves hands-on optimization, contributing to open-source projects, and collaborating with product/sales teams. | ServePost-train | 8 |
| Member of Technical Staff - ML Training Systems The role focuses on engineering for ML training systems, specifically optimizing the training process for production machine learning models, including language models. This involves working with frameworks like PyTorch and high-level training libraries, and optimizing for performance bottlenecks. | Data | 8 |
| Forward Deployed Engineer - ML The Forward Deployed ML Engineer will partner with leading AI companies to help them achieve state-of-the-art performance on demanding AI workloads like LLM serving and model training. This role involves hands-on work with customers to architect and optimize their AI workloads on the Modal platform, contribute to open-source projects, and collaborate with product and sales teams. Requires 2+ years of ML engineering experience, familiarity with serving and training toolchains, and strong communication skills. | ServePost-train | 8 |
| Member of Technical Staff - Product (Backend) Backend engineer for an AI infrastructure company providing GPU access, instant container startups, and storage for training, batch jobs, and low-latency inference. The role involves building modern web applications end-to-end, working across the stack (TypeScript, Python, ClickHouse), and focusing on observability for large-scale AI workloads. | Serve | 7 |
| Customer Engineer Customer Engineer role focused on AI/ML infrastructure, working directly with customers to debug and optimize workloads, while also shipping fixes, features, and automation to improve the core platform. Requires depth in either low-level infrastructure or ML/AI. | ServeData | 7 |