AI Research Scientist - Msl Fair Foundations

Meta Meta · Big Tech · Menlo Park, CA

Research Scientist role focused on designing and developing novel benchmarks and evaluation methodologies for frontier AI capabilities within Meta Superintelligence Labs (MSL). The role involves measuring and understanding AI capabilities, influencing research direction, and collaborating with researchers and technical leadership. Requires a strong publication record and experience in machine learning research, particularly in evaluation and deep learning.

What you'd actually do

  1. Design novel benchmarks and evaluation methodologies for frontier AI capabilities
  2. Contribute to evaluation frameworks that guide research direction and capability development across MSL
  3. Support the scientific vision for evaluation approaches in emerging modalities and novel model capabilities
  4. Partner with cross-functional research teams across product and model training to identify and prioritize gaps in capability through rigorous evaluation
  5. Work on research workstreams that shape the long-term direction of evaluation science at MSL, working independently while also contributing to team goals and organizational priorities

Skills

Required

  • Ph.D. in Computer Science, Machine Learning, or a related technical field
  • 3+ years of experience in machine learning research, with a focus on evaluation, deep learning, or related areas
  • Demonstrated ability to execute on technical research projects from conception to production
  • Effective communication skills and experience collaborating with technical leadership
  • Multiple first-author publications at top-tier peer-reviewed venues (NeurIPS, ICML, ICLR, ACL, EMNLP, or similar) related to language model evaluation, benchmarking, or deep learning
  • Recognized expertise in machine learning evaluation, benchmarking, or capability measurement
  • Track record of research that has substantially influenced the field of deep learning
  • Hands-on experience with language model post-training, RLHF, or related techniques

Nice to have

  • Bachelor's degree in Computer Science, Computer Engineering, relevant technical field, or equivalent practical experience

What the JD emphasized

  • evaluations are the core of AI progress
  • novel evaluations
  • frontier AI systems
  • frontier AI capability measurement
  • scientific validity
  • methodological rigor
  • frontier AI development
  • novel benchmarks
  • evaluation methodologies
  • emerging modalities
  • novel model capabilities
  • rigorous evaluation
  • evaluation science
  • frontier AI
  • language model evaluation
  • benchmarking
  • deep learning
  • machine learning evaluation
  • benchmarking
  • capability measurement
  • deep learning
  • language model post-training
  • RLHF

Other signals

  • evaluations are the core of AI progress
  • provide the technical capabilities to measure and understand the capabilities of our frontier AI systems
  • develop and validate novel evaluations that shape the future of AI capability measurement
  • shape the scientific foundations of frontier AI development