Research Scientist, Hci-multimodality - Interaction Perception, Pico

ByteDance ByteDance · Big Tech · San Jose, CA · R&D

Research Scientist focused on developing computer vision, NLP, and LLM algorithms for next-generation VR intelligent interaction, including input methods, prediction, error correction, and multimodal fusion. The role involves delivering technical innovations, patents, and research translation, with a focus on lightweight models for VR/edge devices.

What you'd actually do

  1. Develop computer vision-driven VR interaction algorithms centered on intelligent input method systems.
  2. Build LLM & NLP algorithms for prediction, error correction and completion.
  3. Research lightweight LLM/NLP and vision-language multimodal fusion for VR.
  4. Deliver technical innovations, patents and research translation.
  5. Provide technical leadership for the algorithm team.

Skills

Required

  • NLP
  • LLM
  • Transformers
  • Sequence Modeling
  • PyTorch
  • TensorFlow
  • Computer Vision
  • Multimodal Background

Nice to have

  • Intelligent Input Methods
  • Chinese Input Methods
  • LLM Fine-tuning
  • Lightweight Model Deployment
  • VR/Edge Devices
  • VR Interaction
  • Scenario Understanding

What the JD emphasized

  • 5+ years NLP/LLM R&D experience
  • 2+ years leading core algorithms
  • Expertise in Transformers, LLMs and sequence modeling
  • Basic computer vision or multimodal background

Other signals

  • LLM
  • NLP
  • Computer Vision
  • Multimodal Fusion
  • VR Interaction