Senior Software Engineer, Machine Learning

AssemblyAI · AI Frontier · Remote · Engineering

Senior ML Engineer to accelerate the AI research-to-production pipeline by building and improving infrastructure for deploying and testing new models, ensuring efficient, scalable, and reliable production inference systems. Requires strong backend engineering, distributed systems, and containerization experience, with a track record of driving projects from concept to delivery.

What you'd actually do

  1. Design and implement tooling that enables researchers to quickly deploy and evaluate new models in production
  2. Design, build, and maintain high-performance, cost-efficient inference pipelines, making architectural decisions about scaling, reliability, and cost trade-offs
  3. Proactively identify and resolve infrastructure bottlenecks, proposing and scoping improvements to iteration speed and production reliability
  4. Develop and maintain user-facing APIs that interact with our ML systems
  5. Implement comprehensive observability solutions to monitor model performance and system health

Skills

Required

  • Strong backend engineering experience with Python
  • Experience building and operating distributed, containerized applications, preferably on AWS
  • Proficiency implementing observability solutions (monitoring, logging, alerting, tracing) for production systems
  • Ability to design and implement resilient, scalable architectures
  • Track record of independently scoping and delivering complex technical projects from problem identification through production deployment
  • Comfort navigating ambiguity and making pragmatic technical decisions when requirements are unclear or evolving

Nice to have

  • MLOps experience, including familiarity with PyTorch and Kubernetes
  • Experience working in fast-paced environments where you owned technical direction for an area and drove projects with minimal oversight.
  • Experience collaborating with remote, globally distributed teams
  • Comfort working across the entire ML lifecycle from model serving to API development
  • Experience in audio-related domains (ASR, TTS, or other domains involving audio processing)
  • Experience with other cloud providers
  • Familiarity with Bazel and monorepos
  • Experience with alternative ML inference frameworks beyond PyTorch
  • Experience with other programming languages
  • Experience mentoring junior engineers or onboarding teammates onto complex systems

What the JD emphasized

  • Strong backend engineering background
  • track record of independently driving projects from concept to delivery
  • Comfort navigating ambiguity and making pragmatic technical decisions when requirements are unclear or evolving

Other signals

  • AI research-to-production pipeline
  • deploy and safely test new models
  • production inference systems remain efficient, scalable, and reliable
  • MLOps practices