Sr. Software Engineer, Siri Speech

Apple Apple · Big Tech · San Francisco Bay Area · Machine Learning and AI

This role is for a Sr. Software Engineer on the Siri Speech team at Apple, focusing on delivering natural language capabilities using ML. The responsibilities include productizing new models for Apple hardware, designing inference systems, building cloud services, and optimizing performance. The role requires experience in iOS development, generative AI agents for coding, and production model inference, with familiarity with large language models preferred.

What you'd actually do

  1. delivering new natural language capabilities through state-of-the-art Machine Learning techniques
  2. explore how technology can lead to new user experiences
  3. learning about new models and figuring out how to productize them to run on all Apple hardware
  4. Design systems to support model inference with optimization, overrides and asset downloads
  5. Build cloud-based services and develop client protocols
  6. Analyze performance and optimize in complex multi-process environments

Skills

Required

  • MS in Computer Science, Computer Engineering, or equivalent degree
  • 8+ years of industry experience
  • iOS application development with Xcode and Swift
  • generative AI agents for coding tasks
  • running model inference in production environments
  • Excellent communication and presentation skills

Nice to have

  • Familiarity with large language models
  • Knowledge of Apple frameworks
  • Python
  • C++
  • Unix
  • server-side software development
  • design of APIs and client/server protocols

What the JD emphasized

  • 8+ years of industry experience
  • Experienced in iOS application development with Xcode and Swift
  • Experienced in using generative AI agents for coding tasks
  • Familiarity with running model inference in production environments

Other signals

  • delight millions of users
  • state-of-the-art Machine Learning techniques
  • productize them to run on all Apple hardware
  • design systems to support model inference with optimization
  • build cloud-based services
  • develop client protocols
  • analyze performance and optimize in complex multi-process environments