Aiml - Machine Learning Research, Multimodal Foundation Models

Apple Apple · Big Tech · Zurich, Switzerland · Machine Learning and AI

Research role focused on building multimodal foundation models, with a focus on image understanding and generation. The role involves developing algorithms, techniques, and systems that push the frontier of deep learning, with opportunities to publish and apply models to Apple products.

What you'd actually do

  1. building multimodal foundation models at Apple, with a particular focus on image understanding and generation
  2. build infrastructure, datasets, and models with fundamental general capabilities
  3. work with a cross-functional partners to apply these models to ship features to millions of users
  4. developing algorithms, techniques, and systems that push the frontier of deep learning
  5. delight millions of users with Apple products powered by foundation models

Skills

Required

  • Deep learning
  • Python
  • JAX
  • PyTorch
  • Tensorflow

Nice to have

  • track record in applying deep learning techniques to shipping products
  • Multi-modal perception for existing products and future hardware platforms
  • On-device intelligence and learning with strong privacy protections
  • Ability to quickly prototype ideas / solutions, perform critical analysis, and use creative approaches for solving complex problems.
  • Ability to collaborate closely with globally distributed teams.

What the JD emphasized

  • publication record required
  • Deep understanding of multi-task, multi-modal machine learning domain
  • Proficient programming skills in Python and one of the deep learning toolkits such as JAX, PyTorch, or Tensorflow
  • PhD, or equivalent practical experience, in Computer Science, or related technical field

Other signals

  • multimodal foundation models
  • image understanding and generation
  • deep learning research
  • apply these models to ship features