Manager, Deep Learning Algorithms

NVIDIA · Semiconductors · Santa Clara, CA

Manager for Deep Learning Algorithms at NVIDIA, focusing on leading engineering efforts for productizing DL models, optimizing inference, and collaborating with research teams to implement and improve algorithms. The role involves managing a team, aligning priorities, and developing the GPU-accelerated DL platform.

What you'd actually do

  1. Plan, schedule, mentor, and lead the execution of projects and activities of the team. Including creating, optimizing, and deploying inference DL workloads.
  2. Collaborate with internal customers to align priorities across business units
  3. Coordinate projects across different geographic locations
  4. Grow and develop a world-class team
  5. Travel to conferences, other sites, or visit customers occasionally

Skills

Required

  • BSc or equivalent experience
  • 8+ overall years related of overall experience
  • 3 years of management/leadership experience
  • Experience leading multiple software engineering projects
  • Strong experience with Large Language Models (LLMs) and Large Visual-Language Models (VLMs)
  • Excellent programming, debugging, performance analysis, and test design skills
  • Great communication

Nice to have

  • Experience with inference of DL models
  • Experience doing performance analysis and tuning
  • Exposure to inference platforms such as TensorRT-LLM, vLLM, and SGLang
  • Project management tools (e.g. JIRA, Microsoft Project)

What the JD emphasized

  • Ability to work in a multifaceted, product-centric environment is required
  • excellent interpersonal skills are also a requirement
  • Strong experience with Large Language Models (LLMs) and Large Visual-Language Models (VLMs)
  • Excellent programming, debugging, performance analysis, and test design skills
  • Great communication

Other signals

  • leading engineering activities related to productizing Deep Learning models
  • implement and improve the latest algorithms
  • large scale LLM/VLM deployment, inference optimization
  • pushing the boundaries of state-of-the-art research and development
  • GPU-accelerated DL platform