Senior Researcher - Efficient AI

Microsoft Microsoft · Big Tech · Bengaluru, KA, IN · Research Sciences

Senior Researcher focused on advancing efficiency across the AI stack for generative AI serving systems, spanning models, ML frameworks, cloud infrastructure, and hardware. The role involves algorithmic and systems optimization for latency, throughput, and cost, with a strong emphasis on driving research ideas through prototyping, validation, and production deployment.

What you'd actually do

  1. Formulate, develop, and evaluate new algorithmic and system-level approaches for end-to-end AI serving, using analytical modeling and large-scale measurement to study token-level latency, tail latency (p95/p99), throughput-per-dollar, cold-start behavior, warm pool strategies, and capacity planning under multi-tenant SLOs and variable sequence lengths.
  2. Design and experimentally evaluate endpoint configuration and execution policies, including batching, routing, and scheduling strategies, tensor and pipeline parallelism, quantization and precision profiles, speculative decoding, and chunked or streaming generation, and drive the most promising approaches through robust rollout and validation into production.
  3. Perform hardware- and kernel-aware optimization by collaborating closely with model, kernel, compiler, and hardware teams to align serving algorithms with attention/KV innovations and accelerator capabilities.
  4. Build and benchmark experimental prototypes and large-scale measurements to validate research ideas and drive them toward production readiness; produce clear technical documentation, design reviews, and operational playbooks.
  5. Publish research results, file patents, and, where appropriate, contribute to open-source systems and serving frameworks.

Skills

Required

  • Doctorate in relevant field OR Master's Degree in relevant field AND 3+ years related research experience OR Bachelor's Degree in relevant field AND 4+ years related research experience OR equivalent experience.
  • Demonstrated expertise in areas of algorithmic optimization, parallel computing, queuing and scheduling theory, and practical request orchestration under strict SLO constraints.
  • Strong understanding of GPU architecture and memory hierarchies.
  • Proficiency in C++ and Python for high-performance systems, with strong code quality and profiling/debugging skills.
  • Proven record of research impact through publications and/or patents, and experience carrying ideas through to systems that operate at scale in real production environments.
  • Ability to meet Microsoft, customer and/or government security screening requirements

Nice to have

  • Deep understanding of transformer inference efficiency techniques such as sharding strategies, attention optimizations, paged KV caches, speculative decoding, LoRA, sequence packing or continuous batching, and quantization.
  • 3+ years of experience with machine learning frameworks (e.g., PyTorch, TensorFlow) and inference serving frameworks (e.g., vLLM, Triton Inference Server, TensorRT-LLM, ONNX Runtime, Ray Serve, DeepSpeed-MII).
  • 3+ years of experience in GPU programming and optimization, with expert knowledge of CUDA, ROCm, Triton, PTX, CUTLASS, or similar GPU programming frameworks.
  • Background in cost and performance modeling, autoscaling, and multi-region deployment or disaster recovery.

What the JD emphasized

  • end-to-end AI serving
  • production
  • hardware- and kernel-aware optimization
  • production readiness
  • production environments

Other signals

  • end-to-end AI serving optimization
  • hardware/software co-design
  • production deployment