Technical Program Manager, Inference Performance

Anthropic Anthropic · AI Frontier · San Francisco, CA · Technical Program Management

Technical Program Manager focused on inference performance and efficiency for AI models, coordinating launches, managing dependencies, and optimizing runtime and accelerator performance across multiple hardware targets.

What you'd actually do

  1. Lead cross-functional initiatives for new infrastructure integration, establishing clear ownership, timelines, and communication channels between teams. Drive end-to-end planning for major infrastructure transitions including platform modernization and new tech adoption.
  2. Partner with engineering teams to identify optimization opportunities, track performance metrics, and prioritize work that unlocks capacity gains. Coordinate across runtime and accelerator layers to ensure efficiency wins ship without compromising reliability.
  3. Drive end-to-end readiness for model and feature launches across multiple hardware platforms. Establish processes for cross-platform validation, manage launch timelines, and ensure smooth handoffs between runtime, accelerator, and downstream teams.
  4. Own and prioritize the inference deployment roadmap, working closely with engineering leadership to prioritize initiatives and manage dependencies. Provide visibility into upcoming changes and their organizational impact.
  5. Build strong relationships across research, engineering, and product teams to understand requirements and constraints. Translate technical complexities into clear updates for leadership and ensure alignment on priorities and timelines.

Skills

Required

  • Technical Program Management
  • ML/AI systems or large-scale distributed systems
  • Inference systems, compilers, or hardware accelerators
  • Stakeholder management
  • Process improvement
  • Infrastructure scaling
  • Deployment governance

Nice to have

  • Experience in fast-paced environments
  • Strategic planning
  • Tactical execution

What the JD emphasized

  • proven success delivering complex infrastructure programs
  • deep technical understanding of inference systems, compilers, or hardware accelerators
  • Excel at creating structure and processes in ambiguous environments
  • strong stakeholder management skills
  • experience with infrastructure scaling initiatives, hardware integrations, or deployment governance
  • passionate about AI infrastructure
  • understand the unique challenges of deploying and scaling large language models

Other signals

  • inference performance
  • model launches
  • infrastructure integration
  • hardware accelerators