ML Compiler Engineer , Aws Neuron, Annapurna Labs

Amazon Amazon · Big Tech · CA, ON +1 · Software Development

The AWS Neuron team is seeking ML Compiler Engineers to optimize deep learning and GenAI workloads on AWS custom ML accelerators (Inferentia/Trainium). This role involves analyzing and optimizing system-level performance across the entire technology stack, from frameworks to runtime, and designing/implementing compiler optimizations. The position requires a passion for performance analysis, distributed systems, and machine learning, with a focus on improving the performance capabilities of the AWS Neuron SDK.

What you'd actually do

  1. Analyze and optimize system-level performance of machine learning models across the entire technology stack, from frameworks to runtime
  2. Conduct detailed performance analysis and profiling of ML workloads, identifying and resolving bottlenecks in large-scale ML systems
  3. Work directly with customers to enable and optimize their ML models on AWS accelerators, understanding their specific requirements and use cases
  4. Design and implement compiler optimizations, transforming manual performance improvements into automated compiler passes
  5. Collaborate across teams to develop innovative optimization techniques that enhance AWS Neuron SDK's performance capabilities

Skills

Required

  • performance analysis
  • distributed systems
  • machine learning
  • compiler optimizations
  • profiling

Nice to have

  • PyTorch
  • TensorFlow
  • JAX
  • high-performance computing

What the JD emphasized

  • AWS Neuron SDK
  • performance

Other signals

  • ML compiler
  • accelerate deep learning and GenAI workloads
  • performance analysis
  • optimization