Modal

Scaling

Data AI · GPU cloud

Currently tracking 7 active AI roles, up 175% versus the prior 4 weeks. Primary focus: Serve · Engineering.

Hiring
7 / 7
Momentum (4w)
+7 +175%
11 opens last 4w · 4 prior 4w
Salary range
Tracked since
Sep '25
last role today
Hiring velocityscroll left for older weeks
1 new role
Feb 12
1 new role
Oct 21
1 new role
Jun 16
1 new role
Aug 18
1 new role
Sep 1
1 new role
15
3 new roles
22
2 new roles
Jan 5
2 new roles
12
1 new role
19
1 new role
26
1 new role
Feb 2
1 new role
16
3 new roles
23
2 new roles
Mar 2
2 new roles
Apr 13
4 new roles
20

Jobs (7)

7 AI · 28 total active
TitleStageFunctionLocationFirst seenAI score
Member of Technical Staff - ML Performance
Seeking an ML Performance Engineer with 5+ years of experience to optimize ML systems for higher throughput and lower latency. The role involves working with inference engines like vLLM or TensorRT, understanding GPU architecture, and improving ML performance at scale.
ServeEngineeringNew York, NY3w ago8
Member of Technical Staff - Agent DX Research
Research role focused on building an evaluation framework for AI coding agents to improve developer experience on the Modal platform. This involves defining quantitative objectives, measuring performance, and translating insights into product improvements, while staying updated on agent advancements and customer use cases.
AgentEval GateResearchNew York, NYMar 38
Forward Deployed Engineer - ML
Forward Deployed ML Engineer to partner with leading AI companies and foundation model labs to help them achieve state-of-the-art performance on demanding workloads like LLM serving, model training (SFT, RLHF), and audio pipelines. This role involves hands-on optimization, contributing to open-source projects, and collaborating with product/sales teams.
ServePost-trainEngineeringStockholm, SwedenMar 38
Member of Technical Staff - ML Training Systems
The role focuses on engineering for ML training systems, specifically optimizing the training process for production machine learning models, including language models. This involves working with frameworks like PyTorch and high-level training libraries, and optimizing for performance bottlenecks.
DataEngineeringNew York, NYFeb 258
Forward Deployed Engineer - ML
The Forward Deployed ML Engineer will partner with leading AI companies to help them achieve state-of-the-art performance on demanding AI workloads like LLM serving and model training. This role involves hands-on work with customers to architect and optimize their AI workloads on the Modal platform, contribute to open-source projects, and collaborate with product and sales teams. Requires 2+ years of ML engineering experience, familiarity with serving and training toolchains, and strong communication skills.
ServePost-trainEngineeringNew York, NYFeb 238
Member of Technical Staff - Product (Backend)
Backend engineer for an AI infrastructure company providing GPU access, instant container startups, and storage for training, batch jobs, and low-latency inference. The role involves building modern web applications end-to-end, working across the stack (TypeScript, Python, ClickHouse), and focusing on observability for large-scale AI workloads.
ServeEngineeringNew York, NYJan 97
Customer Engineer
Customer Engineer role focused on AI/ML infrastructure, working directly with customers to debug and optimize workloads, while also shipping fixes, features, and automation to improve the core platform. Requires depth in either low-level infrastructure or ML/AI.
ServeDataEngineeringNew York, NYSep '257