Staff Software Engineer, Machine Learning Compilers, Edge Tpu

Google Google · Big Tech · Mountain View, CA +2

Staff Software Engineer focused on building ML compilers for EdgeTPU hardware, optimizing ML models for inference, and working on hardware/software co-optimizations. Collaborates with ML model developers and researchers to deploy models on EdgeTPU.

What you'd actually do

  1. Work as part of the EdgeTPU compiler team, building ML compilers for EdgeTPU hardware and analyzing and improving the compiler quality and performance on optimization decisions, correctness and compilation time.
  2. Work with and extend ML authoring frameworks, including JAX, Pytorch to compile ML models for the EdgeTPU.
  3. Work with ML runtime systems to deploy optimized ML models on the EdgeTPU.
  4. Work with EdgeTPU architects to design the Hardware/Software (HW/SW) interface, and co-optimizations between CPU, GPU, and TPU.
  5. Collaborate with ML model developers, researchers, and EdgeTPU hardware/software teams to accelerate the transition from research ideas to user experiences running on the EdgeTPU.

Skills

Required

  • software development
  • software products
  • software design and architecture
  • Machine Learning compilers (optimization, parallelization, etc.)
  • ML infrastructure (e.g., model deployment, model evaluation, etc.)

Nice to have

  • Master’s degree or PhD in Engineering, Computer Science, or a related technical field
  • optimizing ML models for inference
  • compiling for heterogeneous architectures across IPs, including CPU, GPU, and NPUs
  • hardware-software co-design
  • MLIR or Low Level Virtual Machine (LLVM)
  • compiler development, particularly in the context of accelerator-based architectures, vector instruction optimizations, or vectorizing compilers

What the JD emphasized

  • Machine Learning compilers (optimization, parallelization, etc.)
  • ML infrastructure (e.g., model deployment, model evaluation, etc.)
  • optimizing ML models for inference

Other signals

  • ML compilers
  • EdgeTPU hardware
  • ML model optimization
  • inference