Senior Cloud Acceleration Engineer – Dpu & AI Infra

ByteDance ByteDance · Big Tech · Seattle, WA · Infrastructure

Senior Cloud Acceleration Engineer focused on DPU and AI infrastructure, involving software-hardware co-design to optimize distributed training and inference performance. Requires strong C/C++ and Linux systems development, with experience in networking, distributed systems, or AI/ML systems.

What you'd actually do

  1. Design and develop DPU network software with a focus on high performance, low latency, and reliability.
  2. Collaborate with hardware teams to build software-hardware co-design solutions for networking and storage acceleration.
  3. Explore AI/ML infrastructure acceleration, leveraging DPUs, GPUs, and custom hardware to optimize distributed training and inference.
  4. Drive end-to-end performance optimization, from OS kernels and drivers to user-space runtime systems.
  5. Contribute to architecture design, technical proposals, and long-term research directions.

Skills

Required

  • C/C++ development
  • Linux systems development
  • compute architecture
  • network architecture
  • operating systems
  • software-hardware co-design OR distributed systems OR high-performance networking OR AI/ML systems

Nice to have

  • Ph.D. in related fields
  • network virtualization
  • DPDK
  • hardware acceleration
  • FPGA/ASIC/GPU/CUDA
  • NCCL Collectives
  • AI communication patterns
  • parallelization techniques
  • inference kv cache system
  • data preprocessing system

What the JD emphasized

  • software-hardware co-design
  • AI/ML infrastructure acceleration
  • distributed training and inference
  • high performance
  • low latency

Other signals

  • DPU acceleration for AI/ML workloads
  • Optimize distributed training and inference
  • Software-hardware co-design for AI infrastructure