Helix AI Engineer, Pretraining

Figure AI Figure AI · Robotics · HQ · AI - Helix Team

Figure AI is seeking a Helix AI Engineer, Pretraining to build large-scale foundation models for their humanoid robots. This role focuses on advancing pretraining methods across multimodal data (text, vision, robot experience) to enable generalization, reasoning, and adaptability for embodied AI systems. Responsibilities include designing and training models, developing pretraining strategies, exploring architectures, optimizing distributed training pipelines, and collaborating with other AI teams. Requirements include experience training large-scale foundation models, understanding of modern deep learning architectures, and proficiency in Python/PyTorch.

What you'd actually do

  1. Design and train large-scale foundation models across multimodal data (e.g., text, vision, and robot data)
  2. Develop pretraining strategies that improve generalization, reasoning, and transfer to downstream embodied tasks
  3. Explore and implement architectures including transformer-based and emerging foundation model paradigms
  4. Work on scaling laws, dataset mixture design, and training dynamics for frontier models
  5. Build and optimize large-scale distributed training pipelines across multi-node GPU clusters

Skills

Required

  • Python
  • PyTorch
  • Deep learning architectures
  • Transformers
  • Large-scale distributed training
  • Software engineering

Nice to have

  • Frontier foundation models
  • Multimodal pretraining
  • Scaling laws
  • Dataset curation
  • RLHF
  • Reward modeling
  • Alignment methods
  • Embodied AI
  • Robotics
  • Real-world deployment constraints
  • Publication record

What the JD emphasized

  • Experience training large-scale foundation models or working on pretraining for LLMs or multimodal systems
  • Strong experimental rigor and ability to iterate on model design and training strategies
  • Solid software engineering skills and ability to build scalable, reliable systems
  • Ability to operate independently and drive ambiguous, high-impact technical problems

Other signals

  • foundation models
  • pretraining
  • multimodal data
  • robotics
  • large-scale distributed training