Staff AI Research Engineer, Large User Models

Google Google · Big Tech · Mountain View, CA +1

Staff AI Research Engineer focused on architecting the strategy and roadmap for foundation recommender model pre-training. This role involves owning the research agenda, defining and prioritizing experiments for model quality under compute constraints, and partnering with data, ML infrastructure, and engagement teams to shape roadmaps, define training framework requirements, and establish evaluation benchmarks.

What you'd actually do

  1. Define and execute the long-term strategy for foundation recommender model pre-training, encompassing both model architecture evolution and future training methodologies.
  2. Drive a high-velocity research agenda focused on model quality, prioritizing experiments based on compute capacity and researcher bandwidth.
  3. Partner with ML infrastructure teams to architect training frameworks and ensure the technical ecosystem supports the research and release roadmap.
  4. Collaborate with data teams to plan data collection for pre-training, setting the standards for data quality to meet foundational model objectives.
  5. Establish evaluation benchmarks and maintain engaged leaderboards to track progress against baselines and ensure performance.

Skills

Required

  • transformer-based models (e.g., BERT, T5, GPT, ViT), attention mechanisms, and architectural variations
  • research, leading multiple research efforts and influencing research direction related to foundation models, Large Language Models, etc.
  • software development
  • testing, and launching software products
  • software design and architecture

Nice to have

  • Master’s degree or PhD in engineering, computer science, or a related technical field.
  • data structures and algorithms
  • technical leadership role leading project teams and setting technical direction
  • organization involving cross-functional or cross-business projects
  • publications (e.g., NeurIPS, ICML) or open-source contributions in RecSys, NLP, or multimodal systems.

What the JD emphasized

  • foundation recommender model pre-training
  • model quality
  • compute constraints
  • ML infrastructure
  • training framework requirements
  • evaluation benchmarks

Other signals

  • foundation recommender model pre-training
  • model architecture evolution
  • future training methodologies
  • model quality
  • compute constraints
  • ML infrastructure
  • training framework requirements
  • evaluation benchmarks