AI Hire Signal
JobsCompaniesTrendsInsightsWeekly
JobsStrategy timeline
AI Hire Signal

Tracking AI hiring across 200+ US tech companies. Stage, salary, and stack signals on every role — refreshed weekly.

Contact

Browse

JobsCompaniesTrendsInsightsWeekly

Resources

AboutSitemapRobots

Legal

PrivacyTerms
© 2026 AI Hire Signal·Not affiliated with companies shown

Currently tracking 427 active AI roles, up 208% versus the prior 4 weeks. Primary focus: Agent · Engineering. Salary range $65k–$331k (avg $193k).

Hiring
427 / 428
Momentum (4w)
↑+579 +208%
858 opens last 4w · 279 prior 4w
Salary range · avg $193k
$65k–$331k
USD · disclosed roles only
Tracked since
Nov '25
last role today
Hiring velocityscroll left for older weeks
41 new roles
Nov 17
19 new roles
24
18 new roles
Dec 1
29 new roles
8
24 new roles
15
2 new roles
22
4 new roles
29
18 new roles
Jan 5
18 new roles
12
16 new roles
19
25 new roles
26
28 new roles
Feb 2
31 new roles
9
27 new roles
16
44 new roles
23
56 new roles
Mar 2
42 new roles
9
42 new roles
16
61 new roles
23
62 new roles
30
55 new roles
Apr 6
101 new roles
13
97 new roles
20
180 new roles
27
249 new roles
May 4
332 new roles
11

Jobs (82)

427 AI · 1611 total active
FilteredStageServe×FunctionEngineering×Clear all
Show
Active onlyAI only (≥ 7)
Stage
AllData · 42Pretrain · 37Post-train · 52Serve · 108Agent · 205Eval Gate · 17Ship · 117
Function
AllEngineering · 1013Product · 440Research · 158
Country
AllUnited States · 1172Australia · 63Japan · 62Canada · 59United Kingdom · 45Singapore · 22Malaysia · 20South Korea · 17Brazil · 16China · 16Czech Republic · 16Netherlands · 15Switzerland · 15Mexico · 14Taiwan · 14Finland · 13Romania · 13Ireland · 12Denmark · 11Italy · 11Hong Kong · 10France · 9Spain · 8Sweden · 7Vietnam · 7Greece · 6Philippines · 6Costa Rica · 5Norway · 5Poland · 5Thailand · 5Belgium · 4New Zealand · 4Austria · 3Portugal · 2Saudi Arabia · 2South Africa · 2Turkey · 2Bulgaria · 1Egypt · 1Estonia · 1Kenya · 1Puerto Rico · 1
Sort
AI scoreRecentTitle
TitleStageFunctionLocationFirst seenAI score
Member of Technical Staff, Capacity & Efficiency Infrastructure - MAI Superintelligence Team
This role focuses on optimizing and managing the compute infrastructure for training large-scale AI models. The responsibilities include designing and implementing distributed training systems, building telemetry for performance monitoring, profiling and debugging bottlenecks, and driving architectural improvements for efficiency. The role requires strong software engineering skills in Python and C++, deep understanding of GPU architectures, and experience with distributed computing systems and ML workloads.
ServeEngineeringMountain View, CA +28w ago9
Member of Technical Staff, Multimodal Infrastructure - MAI Superintelligence Team
This role focuses on building and maintaining large-scale infrastructure for multimodal generative models, covering the full development cycle from data processing to training, inference, and serving. It involves working with research scientists and product engineers to optimize performance and drive architectural changes for consumer AI products like Copilot.
1–50 of 82← Prev12Next →
ServePost-train
Engineering
Mountain View, CA +4
Mar 3
9
Member of Technical Staff, Software Co-Design AI HPC Systems - MAI Superintelligence Team
This role focuses on the co-design and productionization of next-generation AI systems at datacenter scale, optimizing end-to-end performance and efficiency. It operates at the intersection of models, systems software, networking, storage, and AI hardware, influencing accelerator design, system architectures, and large-scale AI platforms. The role involves analyzing real workloads, developing performance models, and partnering with various teams to drive high-impact ideas into production systems. It also contributes to research and the broader community through publications and open-sourcing.
ServePretrainEngineeringMountain View, CA +3Feb 129
Member of Technical Staff, Software Co-Design AI HPC Systems - MAI Superintelligence Team
This role focuses on the co-design and productionization of next-generation AI systems at datacenter scale, optimizing performance, efficiency, and cost across hardware and software. It involves analyzing workloads, driving architectural decisions, optimizing distributed systems for training and inference, and influencing AI hardware design. The role also includes performance modeling, prototyping, and mentoring.
ServePretrainEngineeringLondon, United Kingdom +2Feb 59
Software Engineer 2
Software Engineer 2 on the Microsoft Azure AI Inference platform team, responsible for the hosting, optimization, and scaling of the inference stack for Azure AI Foundary models, including those from OpenAI and other OSS providers. The role focuses on designing and implementing core inference infrastructure, improving performance and efficiency for LLMs and GenAI models, and scaling the platform to meet growing demand.
ServeEngineeringRedmond, WA +1yesterday8
Principal Software Engineer - Performance
Principal Software Engineer focused on optimizing the performance of AI model inference, particularly LLMs, across various hardware platforms (GPUs, Microsoft silicon). The role involves deep technical work on the AI software stack, from fundamental abstractions to system-level optimizations, aiming to improve efficiency and reduce costs for large-scale AI deployments, including those for Azure OpenAI service.
ServeEngineeringMountain View, CA +11w ago8
Senior Software Engineering
The AI Frameworks team at Microsoft develops software for training and deploying advanced AI models, collaborating with hardware teams and partners on supercomputers and AI accelerators. This role involves developing and evaluating core algorithmic and hardware technologies for large-scale AI model training and inference, working closely with ML researchers and developers, and with OpenAI on Azure OpenAI service models. The position requires hands-on software design and development skills in languages like Python, C/C++, and CUDA, focusing on LLM optimization technologies, model scripting, and kernel languages.
ServePost-trainEngineeringMountain View, CA +22w ago8
Principal Software Engineer
Principal Software Engineer on the AI Frameworks team at Microsoft, responsible for developing and evaluating core algorithmic and hardware technologies for large-scale AI model training and inference on novel hardware. Collaborates with ML researchers, system engineers, and partners to optimize and scale AI models, build validation tools, and perform software development in languages like Python, C/C++, and CUDA.
ServePost-trainEngineeringMountain View, CA +22w ago8
Principal Software Engineer
Principal Software Engineer focused on optimizing GPU inference for large-scale deep learning models (LLMs/SLMs) within Microsoft's AI-native monetization platform, serving ads, shopping, and Copilot.
ServeEngineeringRedmond, WA +14w ago8
Principal Software Engineer, CoreAI
Principal Engineer on the AI Core Infrastructure team, responsible for large-scale GPU management infrastructure and inference/training platforms powering Microsoft's AI workloads. The role involves setting roadmaps, designing backend services, and providing insights for customers to monitor, troubleshoot, and scale AI training workloads on supercomputers. Focus on ML infrastructure, distributed systems, and observability.
ServePost-trainEngineeringRedmond, WA +14w ago8
Principal Software Engineering - AI Frameworks
Principal Software Engineer on the AI Frameworks team at Microsoft, focusing on developing and optimizing software for running AI models across diverse hardware platforms. This includes working on ONNX, ONNX Runtime for high-performance inferencing and training acceleration, and Foundry Local for on-device inference.
ServeEngineeringRedmond, WA +34w ago8
Principal Software Engineer
Principal Software Engineer role focused on building and supporting large-scale GPU management infrastructure and inference/training platforms for AI workloads at Microsoft. The role involves architecting, designing, and developing core AI infrastructure services and compute, storage, and networking subsystems for LLM training, customization, and inference.
ServePost-trainEngineeringBengaluru, KA, IN +15w ago8
Senior Software Engineer, CoreAI Workload Engines
Senior Software Engineer focused on building and optimizing foundational inference engines and APIs for large-scale AI inference across Azure. The role involves improving latency, throughput, availability, and cost for LLMs, working with OpenAI and open-source models, and developing experimentation capabilities for safe and rapid iteration.
ServeEngineeringRedmond, WA +25w ago8
Principal Software Engineer, CoreAI Workload Engines
Principal Software Engineer focused on building and optimizing foundational inference engines and APIs for large-scale AI inference across Azure. The role involves driving production-grade serving improvements for OpenAI and open-source LLMs, focusing on latency, throughput, availability, and cost efficiency. Responsibilities include making hands-on engine changes, building experimentation capabilities, and designing inference serving architectures to support multitenant AI systems at global scale.
ServeEngineeringRedmond, WA +25w ago8
Principal Software Engineer - CoreAI Model Inference & Serving
Principal Software Engineer role focused on building and scaling the AI data-plane for LLM inferencing across Microsoft and Azure. The role involves designing, coding, and shipping core serving systems, smart routing, and request distribution for a wide range of LLMs, aiming for reliability, efficiency, and ultra-low latency.
ServeEngineeringRedmond, WA +27w ago8
Principal Software Engineer, CoreAI
This role focuses on building and optimizing high-performance runtime systems for large-scale LLM inferencing, specifically for OpenAI chat and multimodal AI models. The engineer will be responsible for systems-level optimization, microservice design, and ensuring the latency, throughput, cost, and reliability of AI inference pipelines.
ServeEngineeringRedmond, WA +17w ago8
Member of Technical Staff, AI Systems Engineer - Microsoft Superintelligence
The role focuses on integrating custom AI silicon with AI inference frameworks like SGLang, optimizing LLM inference performance, and developing custom operators. It involves working with hardware accelerators and potentially non-CUDA ecosystems, aiming to improve AI workload efficiency.
ServeEngineeringZürich, ZH, Switzerland7w ago8
Senior AI Software Architect
Senior AI Software Architect role focused on optimizing AI model performance and enablement on Maia hardware, involving PyTorch, quantization, parallelization, and inference pipelines.
ServeEngineeringRedmond, WA +1Jan 138
Member of Technical Staff, AI Networking - MAI Superintelligence Team
This role focuses on designing, scaling, and optimizing high-performance networks for AI training and inference clusters. The engineer will work on the end-to-end networking architecture, from link-layer to fabric-wide systems, connecting thousands of GPUs. Responsibilities include benchmarking, profiling, debugging, and tuning AI workloads, engineering ultra-low-latency networks, and designing congestion-free transport mechanisms. The goal is to build networking systems that directly accelerate Microsoft's frontier AI models and support the development of advanced AI systems.
ServeEngineeringLondon, United Kingdom +2Jan 68
Member of Technical Staff, Developer Experience - MAI Superintelligence Team
This role focuses on building and optimizing the infrastructure and developer experience for large-scale ML model training and inference, specifically for Microsoft's AI assistant, Copilot. The responsibilities include improving CI/CD pipelines, developing training tools, enhancing cloud infrastructure, and managing model hosting systems for inference and data generation. The role aims to accelerate iteration and improve the quality of AI models powering innovative products.
ServeDataEngineeringMountain View, CA +2Dec '258
Member of Technical Staff, LLM Inference - MAI Superintelligence Team
This role focuses on building and maintaining tools and systems for LLM inference, optimizing compute efficiency, and enabling researchers to run models for various tasks. It involves working with inference frameworks, GPU kernel programming, and distributed systems to improve model performance.
ServeEngineeringMountain View, CA +4Nov '258
Senior Software Engineer - AI Frameworks
Senior Software Engineer role focused on optimizing large language model (LLM) deployment on Microsoft's MAIA AI accelerators and GPUs. The role involves building software across the stack, including PyTorch, inference systems (vLLM, SGLang), and performance-critical runtime/kernel components. Responsibilities include architecting tensor computation primitives, extending PyTorch for custom accelerators, improving inference stacks, and optimizing kernels for LLM inference and training workloads.
ServeEngineeringRedmond, WA +2today7
Senior Applied Scientist
This role focuses on building and scaling Azure's next-generation Model Router, which involves optimizing and deploying LLMs at a global scale. The responsibilities include applying advanced concepts to product needs, working with large-scale computing frameworks for model improvement, deploying and evaluating models in production, and monitoring their performance over time. The role also involves collaboration with product groups, mentorship, and documentation.
ServeShipEngineeringHyderabad, TS, INyesterday7
Senior Software Engineer - AI Frameworks
Senior Software Engineer on the AI Frameworks team at Microsoft, focusing on developing AI software that enables running AI models across various devices and hardware. The role involves hands-on software design and development in C/C++ for large-scale model training and inference on novel AI hardware, requiring a strong engineering background and the ability to solve complex technical problems.
ServeEngineeringRedmond, WA +1yesterday7
Applied Scientist II
The Applied Scientist II will design, develop, and ship AI models into the Teams media stack, focusing on real-time conversation products like Teams. This role involves building end-to-end ML systems, from data cleaning and feature engineering to model training, evaluation, and deployment, with an emphasis on optimizing for performance and memory, and updating deployed models based on A/B testing.
ServePost-trainEngineeringRedmond, WA +12d ago7
Principal Software Engineer - Performance Tooling
The Principal Software Engineer - Performance Tooling role focuses on optimizing the performance of AI models, particularly LLMs, across various hardware platforms (GPUs, CPUs) and software layers. This involves benchmarking, debugging, profiling, and optimizing for large-scale training and inference to reduce deployment time and hardware footprint, contributing to the efficiency of AI services like Azure OpenAI.
ServeEngineeringRedmond, WA +22d ago7
Senior Software Engineer - Performance
Senior Software Engineer focused on optimizing the inference performance of large language models (LLMs) like those from OpenAI, running on various hardware including GPUs and custom Microsoft silicon. The role involves benchmarking, debugging, and optimizing performance to enable efficient deployment at scale for major Microsoft products and Azure services.
ServeEngineeringMountain View, CA +16d ago7
Principal Software Engineer
Principal Software Engineer for Microsoft AI's Copilot Discover team, focusing on backend platform for agentic vertical content generation, AI model serving, data ingestion, caching, and serving. Powers content across Microsoft products for over 1B customers. Requires strong distributed systems, cloud storage, and data processing experience.
ServeAgentEngineeringRedmond, WA +12w ago7
Principal Silicon Performance Architect
This role focuses on optimizing the performance of AI inference workloads by exploring micro-architectural innovations and validating end-to-end performance. The Principal Silicon Performance Architect will own performance modeling, analysis, and simulation infrastructure, working closely with chip, system, and software architects to drive data-backed design decisions for improved throughput, latency, and efficiency.
ServeEngineeringRedmond, WA +22w ago7
Senior Software Engineer - Performance Tooling
Senior Software Engineer focused on performance tooling for AI frameworks, enabling large-scale training and inference of LLMs on various hardware. The role involves benchmarking, debugging, profiling, and optimizing performance for models like OpenAI's LLMs, aiming to reduce deployment time and hardware footprint.
ServeEngineeringRedmond, WA +22w ago7
Senior Software Engineer - Azure Translator & Language AI Team
Senior Software Engineer to work on Azure Translator and Language AI services, focusing on building and scaling large-scale AI systems for natural language processing.
ServeEngineeringRedmond, WA +12w ago7
Principal Software Engineer
The Azure AI Inferencing team is seeking a Principal Software Engineer to lead the architecture and design of a large-scale, high-throughput, low-latency model-serving platform for Azure OpenAI generative models, supporting billions of requests daily. The role involves end-to-end ownership of solution quality, cross-team collaboration, incident response, and championing security, privacy, and Responsible AI.
ServeEngineeringBengaluru, KA, IN3w ago7
Principal Software Engineer - CoreAI
Principal Software Engineer to build and scale the core serving systems, request routing, and distribution for all LLMs across Microsoft and Azure customers. The role focuses on delivering inference capabilities reliably, efficiently, and with ultra-low latency for a wide range of AI-powered product experiences.
ServeEngineeringRedmond, WA +13w ago7
Principal Software Engineer
Principal Software Engineer to design and build a Postgres-based database for modern, AI-native, agent-driven workloads within Microsoft Fabric. The role involves innovating on query planning, execution, and storage layers to support high-performance data access for next-generation applications, leveraging open storage formats and engines.
ServeEngineeringBengaluru, KA, IN4w ago7
Member of Technical Staff -Platform Engineering Manager
This role is for a Platform Engineering Manager at Microsoft, focusing on building and scaling the AI platform services that power Copilot, Microsoft's personal AI assistant. The role involves managing a team to develop APIs for finetuning, deployment, and core Copilot experiences, collaborating with AI researchers and product teams to bring AI products to life. The focus is on high-performance, secure, and scalable backend systems for consumer-facing AI products.
ServeEngineeringMountain View, CA +14w ago7
ML - Principal Software Engineer
Principal Software Engineer role focused on building high-performance software for AI capabilities across Windows & Devices. The role involves architecting and building code for deploying ML models at scale, optimizing edge execution, and guiding system-level decisions for inference, memory, power, and security. It requires defining ML infrastructure strategy and has preferred experience in architecting ML inference pipelines for LLMs, local model integrations, and hardware-aware optimizations.
ServeEngineeringHyderabad, TS, IN5w ago7
Member of Technical Staff, Full Stack - ML Efficiency & Observability - MAI Superintelligence Team
Full Stack Engineer on the MAI Superintelligence Team focused on ML Efficiency & Observability, building capacity management portals and visibility into model performance for ML researchers and executives. The role involves designing and developing features for user interfaces, integrating with backend APIs for training frameworks, and contributing to internal tooling and infrastructure.
ServeEngineeringMountain View, CA +26w ago7
Principal AI Network Architect
This role focuses on the network architecture for AI accelerator platforms, specifically for high bandwidth and low latency networks critical for AI GPU clusters. The Principal AI Network Architect will evaluate, design, and optimize the network stack from hardware to software kernels, influencing Azure product roadmaps and working with state-of-the-art networking labs. The role requires deep expertise in networking technologies and familiarity with AI model execution pipelines.
ServeEngineeringRedmond, WA +46w ago7
Principal Software Engineering Manager
Principal Software Engineering Manager to lead a team focused on improving the efficiency, scalability, reliability, and cost of the core infrastructure powering Microsoft 365 Copilot experiences. The role involves acting as a coach, guiding technical design, driving performance optimizations in collaboration with research teams, and ensuring the health and availability of live services.
ServeEngineeringRedmond, WA +16w ago7
Principal Software Engineer
Principal Software Engineer to advance ad-serving infrastructure, focusing on performance, efficiency, and scalability of next-gen model serving and inference platforms for Ads. Designs and optimizes high-performance serving systems and GPU inference frameworks for deep learning and LLM workloads.
ServeEngineeringRedmond, WA +16w ago7
Senior Principal Engineering Manager
Lead and grow a team building and operating world-class research compute infrastructure, including large-scale GPU clusters and agentic development tools, for Microsoft Research globally.
ServeEngineeringRedmond, WA +17w ago7
Senior Software Engineer--Backend--Microsoft Copilot
Senior Software Engineer for Microsoft Copilot's backend platform, focusing on scaling AI services, integrating AI models, and providing tools for engineers. Requires strong backend and cloud infrastructure experience.
ServeEngineeringBeijing, Beijing, China +17w ago7
Senior Software Engineer - CoreAI Model Inference & Serving
Senior Software Engineer role focused on building and scaling the AI data-plane for LLM inferencing across Microsoft and Azure. The role involves designing, coding, and shipping core serving systems, smart routing, and request distribution for a wide range of LLMs, aiming for reliability, efficiency, and ultra-low latency.
ServeEngineeringRedmond, WA +27w ago7
Principal Software Engineering--Backend--Microsoft Copilot
This role focuses on building and scaling the backend platform for Microsoft Copilot, integrating with AI models and empowering Copilot teams. The engineer will design, develop, and maintain performant and secure AI Platform services, ensuring reliability, scalability, and performance. The role requires experience with public cloud infrastructure, containerization, and production software development.
ServeEngineeringBeijing, Beijing, China +18w ago7
MTS - Platform Engineering Manager
This role is for a Platform Engineering Manager at Microsoft AI, focusing on building and scaling the AI platform services that power Copilot. The role involves managing a team to develop secure, performant APIs for finetuning and deploying core AI experiences, collaborating with various teams, and ensuring high-quality code delivery in a fast-paced consumer-facing environment.
ServeEngineeringRedmond, WA +18w ago7
Principal Software Engineer, CoreAI
This role focuses on building and operating the foundational GPU accelerated infrastructure for large-scale AI training and inference across Azure. It involves designing systems for GPU management, scheduling, isolation, and sharing, as well as optimizing performance, reliability, and utilization of GPU fleets. The role also requires driving end-to-end platform features, including observability and diagnostics, and influencing platform architecture.
ServeEngineeringRedmond, WA +28w ago7
Senior Software Engineer--Infra-Microsoft Copilot
The role focuses on building and scaling the backend platform for Microsoft Copilot, including integrations with AI models and tools for engineering teams. The engineer will design, develop, and maintain performant and secure AI Platform services, ensuring reliability, scalability, and performance. This involves working with public cloud infrastructure, containerization technologies, and production software release.
ServeEngineeringBeijing, Beijing, China +18w ago7
Firmware Engineer
Firmware Engineer role focused on designing, developing, and debugging firmware for Azure's custom AI accelerator silicon. This involves working across silicon, hardware, and software teams to enable advanced AI workloads and support data center deployment.
ServeEngineeringRedmond, WA +1Feb 257
Member of Technical Staff, Site Reliability Engineer (HPC) - MAI SuperIntelligence Team
The role is for a Site Reliability Engineer (SRE) focused on High Performance Computing (HPC) infrastructure for AI model training and inference. The engineer will ensure the reliability, availability, and efficiency of large-scale distributed AI systems, including GPU clusters, and will be involved in monitoring, automation, incident management, and security.
ServeEngineeringMountain View, CA +1Feb 177
Member of Technical Staff, HPC Operations Engineering Manager
This role manages a team of Site Reliability Engineers responsible for the reliability and efficiency of large-scale distributed AI infrastructure, specifically for training, fine-tuning, and serving generative AI models. The focus is on leading operations, observability, automation, incident management, and security within hybrid cloud/on-prem CPU+GPU environments, collaborating closely with ML engineers and platform teams.
ServePost-trainEngineeringMountain View, CA +1Feb 117