Research Scientist, Robotics, Embodied Ai, Deepmind

Google Google · Big Tech · Mountain View, CA +1

Research Scientist role focused on developing Vision Language Action (VLA) models and agentic systems for physical robots, enabling them to perceive, plan, think, use tools, and act in real-world environments. The role involves training embodied reasoning VLMs, working with simulators and real robots, and advancing areas like general-purpose robotics, human-robot interaction, and continual learning.

What you'd actually do

  1. Design, implement, train, and evaluate models and algorithms for robotic agents to unlock new robot capabilities.
  2. Write software to implement research ideas and iterate.
  3. Participate in research, including learning from simulation, reinforcement learning, learning from demonstrations, models, transformers, video generation, robot control, and humanoid robots.
  4. Work effectively within collaborative teams despite changing agendas to meet ambitious research goals.
  5. Generate ideas, set up experiments, and test hypotheses. Report and present research findings both internally and externally.

Skills

Required

  • PhD degree in Computer Science, a related field, or equivalent practical experience
  • Experience training embodied reasoning VLMs (Vision Language Models)
  • Experience working with simulators and real-world robots
  • One or more scientific publication submission(s) for conferences, journals, or public repositories (such as NeurIPS, ICML, ICLR)

Nice to have

  • AI technology
  • Python programming
  • robotics

What the JD emphasized

  • PhD degree in Computer Science, a related field, or equivalent practical experience
  • Experience training embodied reasoning VLMs (Vision Language Models)
  • Experience working with simulators and real-world robots
  • One or more scientific publication submission(s) for conferences, journals, or public repositories (such as NeurIPS, ICML, ICLR)

Other signals

  • physical agents
  • robots perceive, plan, think, use tools, and act
  • Vision Language Action (VLA) models
  • Gemini Robotics
  • Gemini Robotics On-Device
  • Gemini agent with spatial understanding
  • robots to perform a range of tasks, respond interactively to their environment, achieve dexterity, and reason over long multi-step tasks
  • general purpose robotics
  • real world understanding
  • action generalization
  • human robot interaction
  • whole-body control
  • continual learning
  • reinforcement learning
  • learning from demonstrations
  • models
  • transformers
  • video generation
  • robot control
  • humanoid robots