AI Research Scientist, Vlm (vision Language Models)

Meta Meta · Big Tech · Bellevue, WA +1

AI Research Scientist focused on Vision-Language Models (VLMs) at Meta, pushing the state-of-the-art in multimodal reasoning and generation. The role involves long-term research, experimental execution, publication, and potential application to Meta's products. Requires a PhD and strong publication record in relevant AI conferences.

What you'd actually do

  1. Push state of the art in multimodal generative AI
  2. Explore new techniques for advanced reasoning and multimodal understanding for AI Assistants
  3. Mentor and work with AI/ML engineers to find a path from research to production

Skills

Required

  • PhD in AI, computer science, or related technical fields
  • Publications in machine learning, computer vision, NLP, speech
  • Experience writing software and executing complex experiments involving large AI models and datasets
  • Python
  • PyTorch

Nice to have

  • Bachelor's degree in Computer Science, Computer Engineering, relevant technical field, or equivalent practical experience
  • Mentor other team members
  • Play a significant role in healthy cross-functional collaboration

What the JD emphasized

  • First (joint) author publications experience at peer-reviewed AI conferences
  • Direct experience in generative AI and LLM research

Other signals

  • multimodal reasoning
  • generative AI
  • LLM research
  • state of the art