AI Hire Signal
JobsCompaniesTrendsInsightsWeekly
JobsStrategy timeline
AI Hire Signal

Tracking AI hiring across 200+ US tech companies. Stage, salary, and stack signals on every role — refreshed weekly.

Contact

Browse

JobsCompaniesTrendsInsightsWeekly

Resources

AboutSitemapRobots

Legal

PrivacyTerms
© 2026 AI Hire Signal·Not affiliated with companies shown

Currently tracking 995 active AI roles, up 64% versus the prior 4 weeks. Primary focus: Agent · Engineering. Salary range $65k–$465k (avg $196k).

Hiring
995 / 995
Momentum (4w)
↑+403 +64%
1033 opens last 4w · 630 prior 4w
Salary range · avg $196k
$65k–$465k
USD · disclosed roles only
Tracked since
Oct '24
last role today
Hiring velocityscroll left for older weeks
2 new roles
Oct 7
1 new role
Feb 3
1 new role
Mar 10
1 new role
17
1 new role
24
2 new roles
31
1 new role
Apr 14
4 new roles
28
2 new roles
May 12
1 new role
19
1 new role
26
3 new roles
Jun 2
1 new role
9
4 new roles
16
2 new roles
23
2 new roles
30
2 new roles
Jul 14
12 new roles
21
3 new roles
28
4 new roles
Aug 4
5 new roles
11
2 new roles
18
3 new roles
25
11 new roles
Sep 1
4 new roles
8
9 new roles
15
4 new roles
22
8 new roles
29
7 new roles
Oct 6
9 new roles
13
8 new roles
20
14 new roles
27
13 new roles
Nov 3
20 new roles
10
14 new roles
17
20 new roles
24
21 new roles
Dec 1
14 new roles
8
19 new roles
15
12 new roles
22
8 new roles
29
29 new roles
Jan 5
22 new roles
12
25 new roles
19
67 new roles
26
64 new roles
Feb 2
71 new roles
9
52 new roles
16
80 new roles
23
110 new roles
Mar 2
135 new roles
9
129 new roles
16
136 new roles
23
136 new roles
30
164 new roles
Apr 6
194 new roles
13
251 new roles
20
237 new roles
27
304 new roles
May 4
241 new roles
11

Jobs (27)

995 AI · 2722 total active
FilteredStageEval Gate×
Show
Active onlyAI only (≥ 7)
Stage
AllData · 240Pretrain · 9Post-train · 93Serve · 179Agent · 537Eval Gate · 27Ship · 367
Function
AllEngineering · 1921Product · 589Research · 209
Country
AllUnited States · 1610Canada · 112United Kingdom · 82Australia · 57Netherlands · 49India · 48Japan · 45China · 34Poland · 31Spain · 27Taiwan · 22Brazil · 20Singapore · 19South Korea · 13Belgium · 12Germany · 12Ireland · 11Hong Kong · 9Romania · 9France · 8Mexico · 8Costa Rica · 7South Africa · 7Philippines · 6Switzerland · 6Sweden · 5Italy · 4New Zealand · 3Thailand · 3Egypt · 2Greece · 2Malaysia · 2Puerto Rico · 2Saudi Arabia · 2Vietnam · 2Austria · 1Denmark · 1Estonia · 1Finland · 1Hungary · 1Norway · 1Turkey · 1
Sort
AI scoreRecentTitle
TitleStageFunctionLocationFirst seenAI score
Applied Scientist II, Alexa Sensitive Content Intelligence (ASCI)
This role focuses on building AI safety systems for Alexa, ensuring LLMs provide safe and trustworthy responses. It involves pioneering solutions in Responsible AI, designing automated testing systems, creating intelligent evaluation systems, building models that understand human values, and crafting AI agents for real-time detection and fixing of production issues. The role emphasizes frontier research with immediate real-world impact, aiming to set industry standards for responsible AI.
Eval GatePost-trainResearchBellevue, WA6w ago9
Applied Science Manager, Sponsored Products and Brands
Manager for a Continuous Model Evaluation and Learning workstream within Amazon Ads' Sponsored Products and Brands team. The role involves leading a team of applied scientists and engineers to build and ship an evaluation and remediation framework for an agentic brand-intelligence system. This includes designing evaluation metrics, developing optimization engines for prompts and synthetic data, and ensuring offline-to-online consistency for quality improvements. The goal is to enable autonomous detect-diagnose-remediate loops to scale quality across brand skills.
Eval GateAgent
Engineering
NY +1
1w ago
8
Data Scientist, AWS Quick Data
The Data Scientist will focus on developing evaluation and benchmarking datasets for generative AI capabilities within the Amazon Quick Suite enterprise AI platform. This includes leveraging LLMs for synthetic data generation, creating ground truth datasets, leading human annotation initiatives, and contributing to Responsible AI efforts to ensure enterprise-readiness, safety, and effectiveness of AI at scale.
Eval GateDataEngineeringSanta Clara, CA1w ago8
Data Scientist, AWS Quick Data
The Data Scientist II will focus on developing evaluation and benchmarking datasets for enterprise AI features, specifically for Amazon Quick Suite. This involves leveraging Generative AI techniques, LLMs for synthetic data generation, and LLM-as-a-judge settings to assess model performance, ensure data quality, and contribute to Responsible AI initiatives. The role also includes building scalable data pipelines and tools for continuous evaluation.
Eval GateDataEngineeringSanta Clara, CA4w ago8
Data Scientist, AWS Quick Data
The Data Scientist will focus on developing evaluation and benchmarking datasets for generative AI capabilities within the Amazon Quick Suite enterprise AI platform. This includes leveraging LLMs for synthetic data generation, creating ground truth datasets, leading human annotation initiatives, and contributing to Responsible AI efforts to ensure enterprise-readiness, safety, and effectiveness of AI at scale.
Eval GateDataEngineeringSanta Clara, CA6w ago8
AI Principal Product Manager-Technical, Alexa Responsible AI
The AI Principal PMT for Alexa Responsible AI will define the standard for how Alexa earns and keeps customer trust. This role owns the product discipline of Responsible AI, defining customer experiences for safety guardrails, trust signals, and evaluation frameworks. The PMT will set product vision and strategy, lead cross-functional alignment across Applied Science, Engineering, Legal, Policy, and UX, and ensure the full responsible product experience including safety, privacy, and security. The role requires technical depth in LLMs and AI safety, understanding how models fail and writing requirements for safety model development and evaluation system design. The PMT will also mentor other PMs and influence Responsible AI scaling across Alexa.
Eval GatePost-trainProductBellevue, WAFeb 258
Manager, Program Management, Alexa Sensitive Content Intelligence (ASCI)
Manager, Program Management for Alexa Sensitive Content Intelligence (ASCI) team, focusing on shaping how Alexa protects customers from harmful content using generative AI and responsible AI guardrails. The role involves strategic leadership, cross-functional program delivery, and team building, with a strong emphasis on data and LLM fluency, defining and executing roadmaps for responsible AI, and ensuring program execution through metrics and mechanisms.
Eval GateAgentProductIN, KA, Bengaluru4d ago7
Applied Scientist, AWS Automated Reasoning
Applied Scientist role focused on automated reasoning, privacy, and sovereignty within AWS. The role involves solving complex problems, designing and implementing solutions, and providing cross-organizational technical influence. Requires a PhD or Master's with significant applied research experience in areas like SAT, SMT, theorem proving, symbolic simulation, program analysis, or type systems. Experience with specific programming languages like O'Caml, Dafny, Haskell, Lean, or Rust is preferred.
Eval GateResearchSeattle, WA1w ago7
Sr. Software Development Engineer, Automated Reasoning Group
Senior Software Development Engineer role focused on applying Automated Reasoning to verify Generative AI outputs, specifically addressing hallucinations within AWS services. The role involves designing and building new services and capabilities at scale, contributing to the evolution of the Automated Reasoning Checks (ARc) service, and making automated reasoning more accessible within AWS.
Eval GateEngineeringNY +12w ago7
AI Benchmarking Lead, Performance Benchmarking Evaluation
This role focuses on ensuring the quality and reliability of AI model evaluations for Amazon's Seller Assistant copilot. The primary responsibilities involve benchmarking AI models, evaluating audits performed by a core auditing team, improving audit consistency, and enforcing quality standards. The goal is to scale AI model evaluation coverage and ensure high-quality outcomes for sellers.
Eval GateEngineeringIN, TS, Hyderabad3w ago7
AI Benchmarking Lead, Performance Benchmarking Evaluation
This role focuses on ensuring the quality and reliability of AI model evaluations for Amazon's Seller Assistant copilot. The primary responsibilities involve benchmarking AI models, evaluating audits performed by a core auditing team, improving audit consistency, and enforcing quality standards. The goal is to scale AI model evaluation coverage and ensure high-quality outcomes for sellers.
Eval GateEngineeringIN, TS, Hyderabad3w ago7
AI Benchmarking Lead, Performance Benchmarking Evaluation
This role focuses on ensuring the quality and reliability of AI model evaluations for Amazon's Seller Assistant copilot. The primary responsibilities involve benchmarking AI models, evaluating audits performed by a core auditing team, improving audit consistency, and enforcing quality standards. The goal is to scale AI model evaluation coverage and ensure high-quality outcomes for sellers.
Eval GateEngineeringIN, TS, Hyderabad3w ago7
AI Benchmarking Lead, Performance Benchmarking Evaluation
This role focuses on ensuring the quality and reliability of AI model evaluations for Amazon's Seller Assistant copilot. The primary responsibilities involve benchmarking AI models, evaluating audits performed by a core auditing team, improving audit consistency, and enforcing quality standards. The goal is to scale AI model evaluation coverage and ensure high-quality outcomes for sellers.
Eval GateEngineeringIN, TS, Hyderabad3w ago7
AI Benchmarking Lead, Performance Benchmarking Evaluation
This role focuses on ensuring the quality and reliability of AI model evaluations for Amazon's Seller Assistant copilot. The primary responsibilities involve benchmarking AI models, evaluating audits performed by a core auditing team, improving audit consistency, and enforcing quality standards. The goal is to scale AI model evaluation coverage and ensure high-quality outcomes for sellers.
Eval GateEngineeringIN, TS, Hyderabad5w ago7
AI Benchmarking Lead, Performance Benchmarking Evaluation
This role focuses on ensuring the quality and reliability of AI model evaluations for Amazon's Seller Assistant copilot. The primary responsibilities involve benchmarking AI models, evaluating audits performed by a core auditing team, improving audit consistency, and enforcing quality standards. The goal is to scale AI model evaluation coverage and ensure high-quality outcomes for sellers.
Eval GateEngineeringIN, TS, Hyderabad5w ago7
AI Benchmarking Lead, Performance Benchmarking Evaluation
This role focuses on ensuring the quality and reliability of AI model evaluations for Amazon's Seller Assistant copilot. The primary responsibilities involve benchmarking AI models, evaluating audits performed by a core auditing team, improving audit consistency, and enforcing quality standards. The goal is to scale AI model evaluation coverage and ensure high-quality outcomes for sellers.
Eval GateEngineeringIN, TS, Hyderabad5w ago7
Senior Applied Scientist, Fauna
Senior Applied Scientist role focused on developing evaluation frameworks and data collection protocols for robotic capabilities. The role involves designing how to measure, stress-test, and improve robot behavior, building infrastructure for teleoperation, evaluation, and learning, and analyzing results to identify performance gaps. It requires expertise in robotics, ML, and human-in-the-loop systems, with a focus on turning capability goals into measurable evaluation systems.
Eval GateAgentEngineeringNY +15w ago7
AI Benchmarking Lead, Performance Benchmarking Evaluation
This role focuses on ensuring the quality and reliability of AI model evaluations for Amazon's Seller Assistant copilot. The primary responsibilities involve benchmarking AI models, evaluating audits performed by a core auditing team, improving audit consistency, and enforcing quality standards. The goal is to scale AI model evaluation coverage and ensure high-quality outcomes for sellers.
Eval GateEngineeringIN, TS, Hyderabad5w ago7
AI Benchmarking Lead, Performance Benchmarking Evaluation
This role focuses on ensuring the quality and reliability of AI model evaluations for Amazon's Seller Assistant copilot. The primary responsibilities involve benchmarking AI models, evaluating audits performed by a core auditing team, improving audit consistency, and enforcing quality standards. The goal is to scale AI model evaluation coverage and ensure high-quality outcomes for sellers.
Eval GateEngineeringIN, TS, Hyderabad5w ago7
AI Benchmarking Lead, Performance Benchmarking Evaluation
This role focuses on ensuring the quality and reliability of AI model evaluations for Amazon's Seller Assistant copilot. The primary responsibilities involve benchmarking AI models, evaluating audits performed by a core auditing team, improving audit consistency, and enforcing quality standards. The goal is to scale AI model evaluation coverage and ensure high-quality outcomes for sellers.
Eval GateEngineeringIN, TS, Hyderabad5w ago7
AI Benchmarking Specialist, SP Support - German, International Seller Growth
This role focuses on evaluating AI systems, specifically LLMs, by designing and executing benchmarking and audit activities. The core responsibilities include assessing model quality, compliance, robustness, and fairness, as well as handling annotations for training and measuring AI models. The role also involves preparing audit reports and ensuring data quality.
Eval GateDataEngineeringIN, KA, Bengaluru5w ago7
AI Benchmarking Lead, Performance Benchmarking Evaluation
The AI Benchmarking Lead will focus on ensuring the quality and reliability of AI model evaluations for Amazon's Seller Assistant copilot. This role involves benchmarking AI models, evaluating audit processes, improving audit consistency, and enforcing quality standards to support the scaling of the product to a wider seller base.
Eval GateEngineeringIN, TS, Hyderabad6w ago7
Software Development Manager, Agentic AI - AgentCore
This role is for a Software Development Manager on the Agentic AI organization's Evaluations & Optimization team at AWS. The manager will lead a team of engineers to build systems for assessing the quality, performance, and reliability of GenAI and agentic systems, as well as optimization solutions. The work involves deep learning, distributed systems, and evaluation science, focusing on building infrastructure and tooling for evaluation workflows.
Eval GateAgentEngineeringSeattle, WA6w ago7
Applied Scientist, Artificial General Intelligence
The Applied Scientist will develop and maintain LLM-as-a-Judge systems and auditing frameworks to ensure the quality of data used for training and evaluating Amazon Nova models, impacting LLM products and services.
Eval GatePost-trainResearchBoston, MA8w ago7
Applied Science Manager, Artificial General Intelligence , Quality Automation
Applied Science Manager for AGI team focusing on quality automation, auditing, and evaluation of LLMs and multimodal systems. Leads a team of scientists to develop quality strategies, auditing frameworks, and research new methodologies to ensure data integrity and model performance. Manages team development, cross-functional communication, and drives research into data impact and utility measurement for AI models.
Eval GatePost-trainEngineeringBellevue, WA8w ago7
AI Benchmarking Specialist, SP Support - Italian, International Seller Growth
This role focuses on evaluating AI systems, specifically LLMs, by designing and executing benchmarking and audit activities. It involves assessing model quality, compliance, robustness, and fairness, with a strong emphasis on handling annotations for training, measuring, and improving AI models. The role also includes preparing audit reports and ensuring data quality based on annotation guidelines.
Eval GateDataEngineeringIN, KA, Bengaluru6w ago6
AI Benchmarking Specialist, SP Support - Spanish, International Seller Growth
This role focuses on evaluating AI systems, specifically LLMs, by designing and executing benchmarking and audit activities. It involves assessing model quality, compliance, robustness, and fairness, as well as handling annotations for training and measuring AI models. The role also includes preparing audit reports and ensuring data quality.
Eval GateDataEngineeringIN, KA, Bengaluru5w ago5