Sr. Research Engineer/scientist (all Levels), Efficient Models

ByteDance ByteDance · Big Tech · Seattle, WA · R&D

Research Engineer/Scientist focused on applied research in Generative AI and CV/Multimodal Understanding, specifically on designing and implementing efficient models for large-scale generative AI through techniques like distillation and compression. The role involves developing methods and infrastructure for transferring capabilities from foundation models into smaller, more efficient models, enabling scalable training, optimization, and deployment, with applications in image generation, video generation, and VLMs.

What you'd actually do

  1. Develop efficient algorithms and architectures for large-scale generative and multimodal models, using techniques such as step distillation, cfg distillation, quantization, and other methods to improve model efficiency (e.g., image generation, video generation, VLM).
  2. Advance scalable generative modeling approaches, including diffusion and autoregressive models, with a focus on acceleration and efficiency.

Skills

Required

  • Expertise in efficient models
  • deep understanding of computational bottlenecks and acceleration methods
  • Proficiency in training generative AI or LLM models
  • PyTorch
  • JAX
  • Strong communication and collaboration skills

Nice to have

  • Ph.D. in GenAI, MLSys or equivalent experience
  • Extensive research experiences in broad GenAI, MLSys, LLM areas
  • image/video generation and editing
  • model compression
  • quantization
  • step/cfg distillation
  • efficient architectures
  • MoE
  • window attention
  • efficient model design
  • reinforcement learning training methods
  • RLHF
  • DPO
  • GRPO

What the JD emphasized

  • distillation
  • compression
  • model efficiency
  • scalable training
  • optimization
  • deployment
  • image generation
  • video generation
  • VLM
  • quantization
  • step distillation
  • cfg distillation
  • model acceleration
  • hardware-efficient inference
  • diffusion
  • autoregressive models
  • RLHF
  • DPO
  • GRPO

Other signals

  • developing methods and infrastructure for transferring capabilities from foundation models into smaller, more efficient models
  • enabling scalable training, optimization, and deployment
  • distillation frameworks, model acceleration, hardware-efficient inference