Research Engineer, Pretraining, Deepmind

Google Google · Big Tech · New York, NY +1

Research Engineer focused on inference-optimized model design for Gemini pretraining, working across the LLM preparation stack (pretraining, fine tuning, serving) and understanding JAX/TPU execution. Requires 5 years of experience in inference-optimized model design and experience with product teams.

What you'd actually do

  1. Focus on inference-optimized model design for Gemini pretraining.
  2. Understand common Accelerated Linear Algebra (XLA) primitives and how JAX code runs on TPUs in practice.
  3. Work across the Large Language Model (LLM) preparation stack (e.g., pretraining, fine tuning, serving) to deliver strong models.

Skills

Required

  • Computer Science
  • Mathematics
  • Applied Stats
  • Machine Learning
  • inference-optimized model design
  • product teams

Nice to have

  • applied research
  • large transformer-based models
  • fine-tuning large models
  • supervised fine-tuning
  • RLHF

What the JD emphasized

  • inference-optimized model design
  • pretraining
  • fine tuning
  • serving

Other signals

  • pretraining
  • LLM
  • inference-optimized model design
  • JAX
  • TPUs