Senior Software Engineer, Real-time AI and Rendering - Holoscan Sdk

NVIDIA · Semiconductors · Santa Clara, CA

Senior Software Engineer at NVIDIA to build the future of real-time AI for sensor-driven applications using the Holoscan Platform. The role involves architecting APIs, prototyping GPU-accelerated algorithms for computer vision, imaging, sensor fusion, and low-latency rendering, and integrating generative models and multimodal foundation models into real-time pipelines. Focus on enabling GPU-resident generative methods for perception, simulation, and robotics.

What you'd actually do

  1. Architect the next generation of Holoscan SDK by developing intuitive, scalable APIs for real-time sensor, imaging, and multimodal data processing—balancing developer usability with peak GPU performance.
  2. Prototype GPU-accelerated algorithms for computer vision, imaging, sensor fusion, and low-latency rendering-translating research into production-grade software.
  3. Build and optimize core GPU libraries for accelerated I/O, streaming, decoding, and visualization, employing CUDA, Vulkan, and GPU-resident data paths.
  4. Contribute to real-time visualization frameworks for medical, robotic, or industrial applications-integrating Vulkan, OpenGL, or Omniverse/RTX-based rendering back-ends.
  5. Benchmark performance rigorously, profiling and optimizing across the full pipeline (Sensor → AI → Render → Display, Sensor → AI → Robotic Control).

Skills

Required

  • C++ (modern standards)
  • Python
  • real-time AI
  • computer vision
  • sensor-driven systems
  • high-performance visualization
  • rendering
  • API design
  • framework development
  • GPU processing
  • rendering pipelines
  • synchronization
  • GPU memory management
  • multi-GPU rendering

Nice to have

  • multimodal foundation models
  • vision-language models
  • streaming workloads
  • CUDA C/C++
  • GPU architecture
  • parallel programming
  • Omniverse Kit
  • GPU rendering frameworks
  • low-latency streaming pipelines
  • multimodal sensor fusion

What the JD emphasized

  • 8+ years of experience building and shipping complex, high-performance imaging, sensor, or rendering software.
  • Familiarity with multimodal or vision-language models and an understanding of how to adapt them to streaming or real-time workloads is a strong plus.
  • Experience adapting VLMs or multimodal foundation models to real-time sensor or video pipelines.

Other signals

  • Generative AI is becoming a central force in real-time sensing, simulation, and robotics
  • Emerging multimodal foundation models (VLMs, video-language models, neural fields) are becoming essential for real-time perception
  • This role will extend Holoscan’s core mission by enabling GPU-resident generative methods that accelerate development, improve simulation fidelity, and unlock new possibilities for real-time perception