AI Hire Signal
JobsCompaniesTrendsInsightsWeekly
JobsStrategy timeline
AI Hire Signal

Tracking AI hiring across 200+ US tech companies. Stage, salary, and stack signals on every role — refreshed weekly.

Contact

Browse

JobsCompaniesTrendsInsightsWeekly

Resources

AboutSitemapRobots

Legal

PrivacyTerms
© 2026 AI Hire Signal·Not affiliated with companies shown

Currently tracking 427 active AI roles, up 208% versus the prior 4 weeks. Primary focus: Agent · Engineering. Salary range $65k–$331k (avg $193k).

Hiring
427 / 428
Momentum (4w)
↑+579 +208%
858 opens last 4w · 279 prior 4w
Salary range · avg $193k
$65k–$331k
USD · disclosed roles only
Tracked since
Nov '25
last role today
Hiring velocityscroll left for older weeks
41 new roles
Nov 17
19 new roles
24
18 new roles
Dec 1
29 new roles
8
24 new roles
15
2 new roles
22
4 new roles
29
18 new roles
Jan 5
18 new roles
12
16 new roles
19
25 new roles
26
28 new roles
Feb 2
31 new roles
9
27 new roles
16
44 new roles
23
56 new roles
Mar 2
42 new roles
9
42 new roles
16
61 new roles
23
62 new roles
30
55 new roles
Apr 6
101 new roles
13
97 new roles
20
180 new roles
27
249 new roles
May 4
332 new roles
11

Jobs (101)

427 AI · 1611 total active
FilteredStageServe×CountryUnited States×Clear all
Show
Active onlyAI only (≥ 7)
Stage
AllData · 42Pretrain · 37Post-train · 52Serve · 108Agent · 205Eval Gate · 17Ship · 117
Function
AllEngineering · 1013Product · 440Research · 158
Country
AllUnited States · 1172Australia · 63Japan · 62Canada · 59United Kingdom · 45Singapore · 22Malaysia · 20South Korea · 17Brazil · 16China · 16Czech Republic · 16Netherlands · 15Switzerland · 15Mexico · 14Taiwan · 14Finland · 13Romania · 13Ireland · 12Denmark · 11Italy · 11Hong Kong · 10France · 9Spain · 8Sweden · 7Vietnam · 7Greece · 6Philippines · 6Costa Rica · 5Norway · 5Poland · 5Thailand · 5Belgium · 4New Zealand · 4Austria · 3Portugal · 2Saudi Arabia · 2South Africa · 2Turkey · 2Bulgaria · 1Egypt · 1Estonia · 1Kenya · 1Puerto Rico · 1
Sort
AI scoreRecentTitle
TitleStageFunctionLocationFirst seenAI score
Principal Researcher
Applied research role focused on advancing efficiency across the AI stack (models, ML frameworks, cloud infrastructure, hardware) for generative AI serving systems. The role involves exploring algorithmic, systems, and hardware/software co-design techniques for optimizations like batching, routing, scheduling, caching, and GPU architecture-aware optimizations. Emphasis on end-to-end ownership, driving research through prototyping, validation, and deployment to production for measurable customer impact.
ServeResearchRedmond, WA +13d ago9
Member of Technical Staff, Capacity & Efficiency Infrastructure - MAI Superintelligence Team
This role focuses on optimizing and managing the compute infrastructure for training large-scale AI models. The responsibilities include designing and implementing distributed training systems, building telemetry for performance monitoring, profiling and debugging bottlenecks, and driving architectural improvements for efficiency. The role requires strong software engineering skills in Python and C++, deep understanding of GPU architectures, and experience with distributed computing systems and ML workloads.
1–50 of 101← Prev123Next →
Serve
Engineering
Mountain View, CA +2
8w ago
9
Member of Technical Staff, Multimodal Infrastructure - MAI Superintelligence Team
This role focuses on building and maintaining large-scale infrastructure for multimodal generative models, covering the full development cycle from data processing to training, inference, and serving. It involves working with research scientists and product engineers to optimize performance and drive architectural changes for consumer AI products like Copilot.
ServePost-trainEngineeringMountain View, CA +4Mar 39
Member of Technical Staff, Software Co-Design AI HPC Systems - MAI Superintelligence Team
This role focuses on the co-design and productionization of next-generation AI systems at datacenter scale, optimizing end-to-end performance and efficiency. It operates at the intersection of models, systems software, networking, storage, and AI hardware, influencing accelerator design, system architectures, and large-scale AI platforms. The role involves analyzing real workloads, developing performance models, and partnering with various teams to drive high-impact ideas into production systems. It also contributes to research and the broader community through publications and open-sourcing.
ServePretrainEngineeringMountain View, CA +3Feb 129
Senior Researcher - AI Systems - Microsoft Research
Senior Researcher in AI Systems at Microsoft Research, focusing on AI infrastructure, ML systems, and HPC systems to improve efficiency and scalability across the full AI model lifecycle, including pre-training, fine-tuning, post-training, and inference.
ServePost-trainResearchRedmond, WA +1Jan 79
Research Intern - LLM Performance Optimization
Research Intern role focused on optimizing the performance of Large Language Models (LLMs), involving architecture and inference performance. Requires PhD student status in a STEM field and experience with LLM architecture or inference performance optimization. Preferred qualifications include experience with GPU kernel performance bottlenecks and optimizing compiler architecture.
ServeResearchRedmond, WA +1Dec '259
Software Engineer 2
Software Engineer 2 on the Microsoft Azure AI Inference platform team, responsible for the hosting, optimization, and scaling of the inference stack for Azure AI Foundary models, including those from OpenAI and other OSS providers. The role focuses on designing and implementing core inference infrastructure, improving performance and efficiency for LLMs and GenAI models, and scaling the platform to meet growing demand.
ServeEngineeringRedmond, WA +1yesterday8
Principal Software Engineer - Performance
Principal Software Engineer focused on optimizing the performance of AI model inference, particularly LLMs, across various hardware platforms (GPUs, Microsoft silicon). The role involves deep technical work on the AI software stack, from fundamental abstractions to system-level optimizations, aiming to improve efficiency and reduce costs for large-scale AI deployments, including those for Azure OpenAI service.
ServeEngineeringMountain View, CA +11w ago8
Senior Software Engineering
The AI Frameworks team at Microsoft develops software for training and deploying advanced AI models, collaborating with hardware teams and partners on supercomputers and AI accelerators. This role involves developing and evaluating core algorithmic and hardware technologies for large-scale AI model training and inference, working closely with ML researchers and developers, and with OpenAI on Azure OpenAI service models. The position requires hands-on software design and development skills in languages like Python, C/C++, and CUDA, focusing on LLM optimization technologies, model scripting, and kernel languages.
ServePost-trainEngineeringMountain View, CA +22w ago8
Principal Software Engineer
Principal Software Engineer on the AI Frameworks team at Microsoft, responsible for developing and evaluating core algorithmic and hardware technologies for large-scale AI model training and inference on novel hardware. Collaborates with ML researchers, system engineers, and partners to optimize and scale AI models, build validation tools, and perform software development in languages like Python, C/C++, and CUDA.
ServePost-trainEngineeringMountain View, CA +22w ago8
Senior Researcher - Efficient AI
Senior Researcher focused on advancing efficiency across the AI stack for generative AI serving systems, spanning models, ML frameworks, cloud infrastructure, and hardware. The role involves algorithmic and systems optimization for latency, throughput, and cost, with a strong emphasis on driving research ideas through prototyping, validation, and production deployment.
ServePost-trainResearchBengaluru, KA, IN3w ago8
Principal Software Engineer
Principal Software Engineer focused on optimizing GPU inference for large-scale deep learning models (LLMs/SLMs) within Microsoft's AI-native monetization platform, serving ads, shopping, and Copilot.
ServeEngineeringRedmond, WA +14w ago8
Principal Product Manager - Foundry Inferencing & Training (CoreAI - multiple roles)
Product Manager for Microsoft Foundry's AI Models & Training organization, focusing on platforms and infrastructure for training, evaluating, launching, and operating AI models at scale. The role involves defining product strategy and roadmaps for model training, inference, experimentation, and platform enablement, driving initiatives across the AI model lifecycle, and enabling internal teams and customers to access and adopt models. Requires strong technical fluency, ability to lead cross-functional initiatives, and experience with AI/ML platforms and infrastructure.
ServePost-trainProductRedmond, WA +44w ago8
Principal Software Engineer, CoreAI
Principal Engineer on the AI Core Infrastructure team, responsible for large-scale GPU management infrastructure and inference/training platforms powering Microsoft's AI workloads. The role involves setting roadmaps, designing backend services, and providing insights for customers to monitor, troubleshoot, and scale AI training workloads on supercomputers. Focus on ML infrastructure, distributed systems, and observability.
ServePost-trainEngineeringRedmond, WA +14w ago8
Principal Software Engineering - AI Frameworks
Principal Software Engineer on the AI Frameworks team at Microsoft, focusing on developing and optimizing software for running AI models across diverse hardware platforms. This includes working on ONNX, ONNX Runtime for high-performance inferencing and training acceleration, and Foundry Local for on-device inference.
ServeEngineeringRedmond, WA +34w ago8
Principal Software Engineer
Principal Software Engineer role focused on building and supporting large-scale GPU management infrastructure and inference/training platforms for AI workloads at Microsoft. The role involves architecting, designing, and developing core AI infrastructure services and compute, storage, and networking subsystems for LLM training, customization, and inference.
ServePost-trainEngineeringBengaluru, KA, IN +15w ago8
Senior Software Engineer, CoreAI Workload Engines
Senior Software Engineer focused on building and optimizing foundational inference engines and APIs for large-scale AI inference across Azure. The role involves improving latency, throughput, availability, and cost for LLMs, working with OpenAI and open-source models, and developing experimentation capabilities for safe and rapid iteration.
ServeEngineeringRedmond, WA +25w ago8
Principal Software Engineer, CoreAI Workload Engines
Principal Software Engineer focused on building and optimizing foundational inference engines and APIs for large-scale AI inference across Azure. The role involves driving production-grade serving improvements for OpenAI and open-source LLMs, focusing on latency, throughput, availability, and cost efficiency. Responsibilities include making hands-on engine changes, building experimentation capabilities, and designing inference serving architectures to support multitenant AI systems at global scale.
ServeEngineeringRedmond, WA +25w ago8
Principal Software Engineer - CoreAI Model Inference & Serving
Principal Software Engineer role focused on building and scaling the AI data-plane for LLM inferencing across Microsoft and Azure. The role involves designing, coding, and shipping core serving systems, smart routing, and request distribution for a wide range of LLMs, aiming for reliability, efficiency, and ultra-low latency.
ServeEngineeringRedmond, WA +27w ago8
Principal Software Engineer, CoreAI
This role focuses on building and optimizing high-performance runtime systems for large-scale LLM inferencing, specifically for OpenAI chat and multimodal AI models. The engineer will be responsible for systems-level optimization, microservice design, and ensuring the latency, throughput, cost, and reliability of AI inference pipelines.
ServeEngineeringRedmond, WA +17w ago8
Senior Researcher - GPU Performance
Applied Research role focused on hardware/software codesign for GPU kernel optimizations to improve efficiency of Large Language Models and Generative AI inference. Involves designing, implementing, and optimizing GPU kernels, researching novel optimization techniques, and profiling performance.
ServeResearchRedmond, WA +1Jan 278
Senior AI Software Architect
Senior AI Software Architect role focused on optimizing AI model performance and enablement on Maia hardware, involving PyTorch, quantization, parallelization, and inference pipelines.
ServeEngineeringRedmond, WA +1Jan 138
Member of Technical Staff, Developer Experience - MAI Superintelligence Team
This role focuses on building and optimizing the infrastructure and developer experience for large-scale ML model training and inference, specifically for Microsoft's AI assistant, Copilot. The responsibilities include improving CI/CD pipelines, developing training tools, enhancing cloud infrastructure, and managing model hosting systems for inference and data generation. The role aims to accelerate iteration and improve the quality of AI models powering innovative products.
ServeDataEngineeringMountain View, CA +2Dec '258
Research Intern - AI/ML Numerics & Efficiency
Research Intern role focusing on ML systems, numeric precision, data types, and compute technologies for AI workloads at Azure scale. The role involves investigating model efficiency through low-precision formats, quantization, ML kernel development, and benchmarking. It aims to inform decisions on compute platforms, acceleration strategies, and system-level optimizations for training and inference of large-scale models.
ServeResearchRedmond, WA +1Dec '258
Member of Technical Staff, LLM Inference - MAI Superintelligence Team
This role focuses on building and maintaining tools and systems for LLM inference, optimizing compute efficiency, and enabling researchers to run models for various tasks. It involves working with inference frameworks, GPU kernel programming, and distributed systems to improve model performance.
ServeEngineeringMountain View, CA +4Nov '258
Research Intern - AI Systems & Architecture
Research internship focused on AI systems and architecture, investigating performance modeling, architectural analysis, and emerging inference mechanisms for large-scale AI workloads. The role involves analyzing hardware, software, and model interactions, developing performance models, and prototyping new inference techniques.
ServeResearchMountain View, CA +2Nov '258
Research Intern - LLM Acceleration
Research intern focused on accelerating large language models (LLMs) by optimizing performance on custom architectures, involving computer architecture and parallel programming.
ServeResearchMountain View, CA +2Nov '258
Research Intern - Systems For Efficient AI
Research intern focused on optimizing AI inference systems, including LLM inference, KV caching, request scheduling, and GPU orchestration, to improve latency, throughput, and cost-efficiency.
ServeResearchRedmond, WA +1Nov '258
Senior Software Engineer - AI Frameworks
Senior Software Engineer role focused on optimizing large language model (LLM) deployment on Microsoft's MAIA AI accelerators and GPUs. The role involves building software across the stack, including PyTorch, inference systems (vLLM, SGLang), and performance-critical runtime/kernel components. Responsibilities include architecting tensor computation primitives, extending PyTorch for custom accelerators, improving inference stacks, and optimizing kernels for LLM inference and training workloads.
ServeEngineeringRedmond, WA +2today7
Senior Applied Scientist
This role focuses on building and scaling Azure's next-generation Model Router, which involves optimizing and deploying LLMs at a global scale. The responsibilities include applying advanced concepts to product needs, working with large-scale computing frameworks for model improvement, deploying and evaluating models in production, and monitoring their performance over time. The role also involves collaboration with product groups, mentorship, and documentation.
ServeShipEngineeringHyderabad, TS, INyesterday7
Senior Software Engineer - AI Frameworks
Senior Software Engineer on the AI Frameworks team at Microsoft, focusing on developing AI software that enables running AI models across various devices and hardware. The role involves hands-on software design and development in C/C++ for large-scale model training and inference on novel AI hardware, requiring a strong engineering background and the ability to solve complex technical problems.
ServeEngineeringRedmond, WA +1yesterday7
Applied Scientist II
The Applied Scientist II will design, develop, and ship AI models into the Teams media stack, focusing on real-time conversation products like Teams. This role involves building end-to-end ML systems, from data cleaning and feature engineering to model training, evaluation, and deployment, with an emphasis on optimizing for performance and memory, and updating deployed models based on A/B testing.
ServePost-trainEngineeringRedmond, WA +12d ago7
Principal Software Engineer - Performance Tooling
The Principal Software Engineer - Performance Tooling role focuses on optimizing the performance of AI models, particularly LLMs, across various hardware platforms (GPUs, CPUs) and software layers. This involves benchmarking, debugging, profiling, and optimizing for large-scale training and inference to reduce deployment time and hardware footprint, contributing to the efficiency of AI services like Azure OpenAI.
ServeEngineeringRedmond, WA +22d ago7
Senior Software Engineer - Performance
Senior Software Engineer focused on optimizing the inference performance of large language models (LLMs) like those from OpenAI, running on various hardware including GPUs and custom Microsoft silicon. The role involves benchmarking, debugging, and optimizing performance to enable efficient deployment at scale for major Microsoft products and Azure services.
ServeEngineeringMountain View, CA +16d ago7
Principal Software Engineer
Principal Software Engineer for Microsoft AI's Copilot Discover team, focusing on backend platform for agentic vertical content generation, AI model serving, data ingestion, caching, and serving. Powers content across Microsoft products for over 1B customers. Requires strong distributed systems, cloud storage, and data processing experience.
ServeAgentEngineeringRedmond, WA +12w ago7
Principal Silicon Performance Architect
This role focuses on optimizing the performance of AI inference workloads by exploring micro-architectural innovations and validating end-to-end performance. The Principal Silicon Performance Architect will own performance modeling, analysis, and simulation infrastructure, working closely with chip, system, and software architects to drive data-backed design decisions for improved throughput, latency, and efficiency.
ServeEngineeringRedmond, WA +22w ago7
Senior Software Engineer - Performance Tooling
Senior Software Engineer focused on performance tooling for AI frameworks, enabling large-scale training and inference of LLMs on various hardware. The role involves benchmarking, debugging, profiling, and optimizing performance for models like OpenAI's LLMs, aiming to reduce deployment time and hardware footprint.
ServeEngineeringRedmond, WA +22w ago7
Senior Software Engineer - Azure Translator & Language AI Team
Senior Software Engineer to work on Azure Translator and Language AI services, focusing on building and scaling large-scale AI systems for natural language processing.
ServeEngineeringRedmond, WA +12w ago7
Principal Software Engineer
The Azure AI Inferencing team is seeking a Principal Software Engineer to lead the architecture and design of a large-scale, high-throughput, low-latency model-serving platform for Azure OpenAI generative models, supporting billions of requests daily. The role involves end-to-end ownership of solution quality, cross-team collaboration, incident response, and championing security, privacy, and Responsible AI.
ServeEngineeringBengaluru, KA, IN3w ago7
Principal Software Engineer - CoreAI
Principal Software Engineer to build and scale the core serving systems, request routing, and distribution for all LLMs across Microsoft and Azure customers. The role focuses on delivering inference capabilities reliably, efficiently, and with ultra-low latency for a wide range of AI-powered product experiences.
ServeEngineeringRedmond, WA +13w ago7
Senior Product Manager - Foundry Inferencing & Training (CoreAI - multiple roles)
Senior Product Manager for Microsoft's Foundry Inference & Training team, focusing on product strategy and execution for AI model platforms. The role involves owning product strategy for AI model training, inference, experimentation, and platform enablement, evolving model offerings, driving developer-facing experiences, and defining efficiency metrics. Collaboration with engineering, data science, finance, and go-to-market teams is key, with a focus on solutions for regulated environments.
ServeProductRedmond, WA +44w ago7
Principal Software Engineer
Principal Software Engineer to design and build a Postgres-based database for modern, AI-native, agent-driven workloads within Microsoft Fabric. The role involves innovating on query planning, execution, and storage layers to support high-performance data access for next-generation applications, leveraging open storage formats and engines.
ServeEngineeringBengaluru, KA, IN4w ago7
Member of Technical Staff -Platform Engineering Manager
This role is for a Platform Engineering Manager at Microsoft, focusing on building and scaling the AI platform services that power Copilot, Microsoft's personal AI assistant. The role involves managing a team to develop APIs for finetuning, deployment, and core Copilot experiences, collaborating with AI researchers and product teams to bring AI products to life. The focus is on high-performance, secure, and scalable backend systems for consumer-facing AI products.
ServeEngineeringMountain View, CA +14w ago7
ML - Principal Software Engineer
Principal Software Engineer role focused on building high-performance software for AI capabilities across Windows & Devices. The role involves architecting and building code for deploying ML models at scale, optimizing edge execution, and guiding system-level decisions for inference, memory, power, and security. It requires defining ML infrastructure strategy and has preferred experience in architecting ML inference pipelines for LLMs, local model integrations, and hardware-aware optimizations.
ServeEngineeringHyderabad, TS, IN5w ago7
Member of Technical Staff, Full Stack - ML Efficiency & Observability - MAI Superintelligence Team
Full Stack Engineer on the MAI Superintelligence Team focused on ML Efficiency & Observability, building capacity management portals and visibility into model performance for ML researchers and executives. The role involves designing and developing features for user interfaces, integrating with backend APIs for training frameworks, and contributing to internal tooling and infrastructure.
ServeEngineeringMountain View, CA +26w ago7
Principal AI Network Architect
This role focuses on the network architecture for AI accelerator platforms, specifically for high bandwidth and low latency networks critical for AI GPU clusters. The Principal AI Network Architect will evaluate, design, and optimize the network stack from hardware to software kernels, influencing Azure product roadmaps and working with state-of-the-art networking labs. The role requires deep expertise in networking technologies and familiarity with AI model execution pipelines.
ServeEngineeringRedmond, WA +46w ago7
Principal Software Engineering Manager
Principal Software Engineering Manager to lead a team focused on improving the efficiency, scalability, reliability, and cost of the core infrastructure powering Microsoft 365 Copilot experiences. The role involves acting as a coach, guiding technical design, driving performance optimizations in collaboration with research teams, and ensuring the health and availability of live services.
ServeEngineeringRedmond, WA +16w ago7
Principal Software Engineer
Principal Software Engineer to advance ad-serving infrastructure, focusing on performance, efficiency, and scalability of next-gen model serving and inference platforms for Ads. Designs and optimizes high-performance serving systems and GPU inference frameworks for deep learning and LLM workloads.
ServeEngineeringRedmond, WA +16w ago7
Senior Principal Engineering Manager
Lead and grow a team building and operating world-class research compute infrastructure, including large-scale GPU clusters and agentic development tools, for Microsoft Research globally.
ServeEngineeringRedmond, WA +17w ago7
Principal Product Manager/Architect - Foundry Inference Platform (CoreAI)
The Principal Product Manager/Architect will define and guide the technical architecture of Microsoft Foundry, an AI inferencing platform focused on reliability, scalability, and efficiency for large-scale GPU fleets. The role involves setting product direction for reliability, GPU fleet efficiency, capacity management, and engaging with strategic customers. Success metrics include platform reliability, GPU utilization, and customer outcomes.
ServeProductRedmond, WA +17w ago7