GPU Computing Engineer - Autonomous Driving

NVIDIA NVIDIA · Semiconductors · Shanghai, China

NVIDIA is seeking a GPU Computing Engineer in Shanghai to analyze Deep Learning models and investigate TensorRT stability and performance issues. The role involves working with a global team on CUDA and TensorRT development, extracting feature requirements, and generating documentation. Requires strong C/C++/Python skills, knowledge of inference networks, and experience with deep learning frameworks like PyTorch.

What you'd actually do

  1. Analyze Deep Learning models and investigate TensorRT stability and performance issues reported by customers or internal teams.
  2. Work with internationally distributed team with remote locations in US, APAC and India for CUDA and TensorRT developing.
  3. Extract the feature requirement or FAQ from the analysis and development and generate the documents.

Skills

Required

  • C++
  • Python
  • Deep learning frameworks (Torch, PyTorch)
  • Inference networks and layers
  • CUDA
  • TensorRT
  • English and Mandarin communication

Nice to have

  • Master Degree
  • Pytorch expertise
  • Customer communication skills

What the JD emphasized

  • Bachelor or equivalent experience of Computer Science or Electrical Engineering is required
  • 3-5+ years of related work
  • Strong programming skills in C and C++ and python
  • Have knowledge about the popular inference network and layers
  • Experience working with deep learning frameworks like Torch and Pytorch
  • Strong written and verbal communications in both English and Mandarin
  • Strong customer communication skills, powerfully motivated to provide highly responsive support as needed.

Other signals

  • GPU computing
  • Deep Learning models
  • TensorRT
  • CUDA