Software Engineer, Core AI Compiler & Runtime

Tesla Tesla · Auto · Palo Alto, CA · Tesla AI

Software Engineer role focused on designing and developing the AI inference stack, including compilers and runtimes, for neural networks powering Tesla's vehicles and Optimus robot. The role involves optimizing performance on custom hardware and collaborating with AI and hardware engineers.

What you'd actually do

  1. Write, debug and maintain robust software for Tesla AI inference (Compiler / Runtime) across FSD, Optimus, and Data Center use-cases
  2. Provide AI developers access to performance-critical hardware features through new APIs and Domain-Specific Languages (DSLs)
  3. Design APIs, compiler and runtime features enabling distributed inference on Tesla’s hardware
  4. Support the integration of our deployment stack with ML frameworks (PyTorch, JAX)
  5. Keep up-to-date and collaborate with ML/compiler open-source community

Skills

Required

  • ML compilers/runtimes (e.g. MLIR, LLVM, XLA, PJRT, TensorRT)
  • Domain-Specific Languages (DSLs) like Triton, cuTile, Pallas
  • CPUs, GPUs and modern AI accelerators
  • computer architecture
  • distributed systems
  • networking
  • collectives
  • ML frameworks (PyTorch, JAX)
  • framework internals
  • C/C++ programming
  • Python
  • modern ML architectures

Nice to have

  • MLIR
  • LLVM
  • XLA
  • PJRT
  • TensorRT
  • Triton
  • cuTile
  • Pallas
  • PyTorch
  • JAX

What the JD emphasized

  • AI inference stack
  • compiler/runtime
  • performance

Other signals

  • AI inference stack
  • compiler/runtime
  • performance optimization
  • MLIR
  • Tesla vehicles
  • Optimus robot