Week 2025-W12
4 new AI roles opened across 1 companies. Highest-signal roles first.
Anthropic· 4 roles
- TPU Kernel Engineer Serve · Engineering 9This role focuses on optimizing ML systems, particularly for TPUs, by designing and implementing kernels to improve performance for research, training, and inference. It involves low-level optimization and providing feedback on model performance impacts.
- Staff Software Engineer, AI Reliability Engineering Serve · Engineering 8Staff Software Engineer focused on AI Reliability Engineering, responsible for defining and achieving reliability metrics for Anthropic's AI systems, including LLM serving and training infrastructure. The role involves designing monitoring, high-availability serving systems, automated failover, incident response, and cost optimization for large-scale AI infrastructure.
- Software Engineer, Inference Scalability and Capability Serve · Engineering 8Software Engineer focused on building and scaling inference systems for LLMs, optimizing performance, reliability, and compute efficiency. This role involves tackling complex distributed systems challenges across the inference stack, from request routing to caching, and supporting new model architectures and inference features.
- Staff Software Engineer, AI Reliability Engineering Serve · Engineering 7Staff Software Engineer focused on AI Reliability Engineering, responsible for defining and achieving reliability metrics for Anthropic's AI products and services, including LLM serving and training systems. This involves designing monitoring, implementing high-availability infrastructure, leading incident response, and optimizing costs for large-scale AI infrastructure.