Senior Research Scientist/engineer - AI Infrastructure

ByteDance ByteDance · Big Tech · San Jose, CA · Algorithm

Seeking an experienced Research Scientist/Engineer to design and build next-generation AI infrastructure at ByteDance, focusing on large-scale systems, AI, and emerging hardware to enable efficient and scalable AI workloads. The role involves architecting the end-to-end AI factory, exploring emerging trends, optimizing ML stack performance, and aligning cross-functional teams.

What you'd actually do

  1. Design and evaluate scalable architectures across the full AI factory — compute, storage, networking, chips, power, and the data and application layers — for large-scale training, RL, and inference workloads. Develop technical proposals for supply-chain and energy constraints alongside silicon and software trade-offs.
  2. Track emerging trends across AI systems, distributed training and RL, and hardware acceleration, as well as adjacent fields such as cognitive science and psychology that inform AI memory and reasoning substrates. Build prototypes and share insights through technical reports.
  3. Analyze and optimize performance across the ML stack — scheduling, networking, storage, training and RL frameworks, and emerging AI memory systems for long-horizon agents — through benchmarking and bottleneck analysis.
  4. Work across research, engineering, hardware, data-center, and product teams to translate AI workload requirements into scalable solutions and drive cross-team initiatives spanning the full AI factory.

Skills

Required

  • PhD in Computer Science, Computer Engineering, Electrical Engineering, or related technical discipline
  • Backgrounds in cognitive science, computational neuroscience, or psychology with strong systems fundamentals
  • Experience in distributed systems, infrastructure engineering, or ML systems
  • Exposure to large-scale training or RL pipelines
  • Comfort evaluating trade-offs across hardware, software, algorithms, energy, and supply-chain constraints
  • Strong proficiency in integrating AI tools into knowledge discovery and research workflows
  • Demonstrated ability to learn quickly and stay productive on a fast-evolving technical horizon
  • Excellent communication skills

Nice to have

  • Experience with large-scale model training and inference
  • distributed pretraining
  • post-training
  • RL
  • KV cache–aware serving
  • GPU/accelerator optimization
  • high-performance networking (e.g., RDMA, NCCL)
  • Experience with heterogeneous AI compute systems
  • large-scale training clusters
  • HPC-style distributed workloads
  • data pipelines for training and evaluation
  • Familiarity with AI memory systems
  • retrieval-augmented architectures
  • agent long-term memory designs
  • Exposure to cognitive-science or psychology literature on memory and reasoning
  • Exposure to chip-level design
  • data-center energy and cooling
  • AI hardware supply-chain considerations
  • Publications in systems and/or machine learning conferences (e.g., NeurIPS, OSDI, SOSP, ASPLOS, MLSys)
  • Contributions to open-source projects

What the JD emphasized

  • AI Factory Architecture
  • Research & Technology Exploration
  • AI Memory & System Performance Optimization
  • Cross-Team Technical Alignment
  • large-scale training
  • RL pipelines
  • large-scale model training
  • distributed pretraining
  • post-training
  • RL
  • KV cache–aware serving
  • GPU/accelerator optimization
  • heterogeneous AI compute systems
  • large-scale training clusters
  • HPC-style distributed workloads
  • data pipelines for training and evaluation
  • AI memory systems
  • retrieval-augmented architectures
  • agent long-term memory designs
  • chip-level design
  • data-center energy and cooling
  • AI hardware supply-chain considerations
  • AI factory

Other signals

  • AI infrastructure
  • large-scale systems
  • emerging hardware
  • AI workloads
  • AI factory