Applied Machine Learning Research Scientist

Cerebras · Semiconductors · Headquarters +2 · AppliedML

This role focuses on applying and scaling modern machine learning techniques, particularly LLM post-training (RLHF, GRPO), on Cerebras' wafer-scale AI chip. The scientist will build and maintain training pipelines, evaluation frameworks, and optimize ML workflows across pretraining, fine-tuning, and alignment stages, working with large datasets and contributing to shared ML infrastructure.

What you'd actually do

  1. Apply post-training techniques (e.g. RLVR, RLHF, GRPO etc.) techniques to improve model performance.
  2. Build and maintain evaluation pipelines to measure model performance across tasks and domains.
  3. Debug issues across the ML stack, including data pipelines, training jobs, model outputs and mixed or lower precision computation.
  4. Collaborate with researchers to translate ML ideas into efficient, scalable implementation.
  5. Design, implement, and scale ML pipelines across all stages of LLM development (pretraining, fine-tuning, alignment).

Skills

Required

  • Python
  • PyTorch
  • machine learning fundamentals
  • deep learning architectures
  • transformers
  • read and understand modern ML papers and implement key ideas

Nice to have

  • large language models (training, fine-tuning, and evaluation)
  • reinforcement learning concepts
  • distributed training frameworks (e.g., FSDP, Megatron)
  • large-scale datasets and data pipelines
  • debugging or optimizing ML systems for performance
  • Contributions to meaningful codebases, projects, or open-source systems

What the JD emphasized

  • implement and improve workflows for LLM pretraining, fine-tuning, and reinforcement learning-based post-training
  • building training pipelines
  • improving model quality
  • iterating on data and evaluation strategies
  • translate cutting-edge ML ideas into reliable, production-ready systems

Other signals

  • implement and improve workflows for LLM pretraining, fine-tuning, and reinforcement learning-based post-training
  • building training pipelines
  • improving model quality
  • iterating on data and evaluation strategies
  • translate cutting-edge ML ideas into reliable, production-ready systems