Enterprise · Enterprise search
Currently tracking 43 active AI roles, with 26 new openings in the last 4 weeks. Primary focus: Agent · Engineering. Salary range $120k–$300k (avg $210k).
| Title | Stage | AI score |
|---|---|---|
| Machine Learning Engineer, LLM Evals & Observability Machine Learning Engineer focused on LLM Evals & Observability for Glean's Work AI platform, responsible for designing evaluation datasets, building large-scale evaluation pipelines, developing LLM-powered judges, and creating observability infrastructure for AI agents. | Eval GateAgent | 8 |
| Machine Learning Engineer, LLM Evals & Observability Machine Learning Engineer focused on LLM Evals & Observability for an enterprise AI platform. The role involves designing evaluation datasets, building large-scale evaluation pipelines, creating LLM-powered judges, evaluating new models before launch, and building observability infrastructure for AI agents. The goal is to ensure the reliability and quality of Glean's AI Assistant and Agents. | Eval GateAgent |
| 8 |
| Product Manager, AI Quality Product Manager for Glean Model Hub, responsible for evaluating LLM models, defining the roadmap for Glean's LLM portfolio, managing provider relationships, and setting product strategy for customer LLM experimentation. This role involves forecasting LLM usage, costs, and capacity, and driving customer adoption of new LLMs and AI capabilities. | Eval GateAgent | 7 |
| Product Manager, AI Quality Product Manager for Glean Model Hub, responsible for evaluating LLM models, defining the roadmap for Glean's LLM portfolio, managing provider relationships, and setting product strategy for customer LLM experimentation. The role involves owning LLM usage projections, cost, and capacity planning, while driving customer enablement and adoption of LLM-native capabilities. The company is an AI platform for enterprise search, AI assistants, and AI agents. | Eval Gate | 7 |
| Senior/Staff Applied Scientist Senior/Staff Applied Scientist at Glean, an enterprise Work AI platform. The role focuses on developing and maintaining an A/B experimentation platform, conducting end-to-end evaluations of AI use cases (including summarization, reasoning, tool use, coding), and designing best practices for evaluation problems. The role requires strong statistics/ML skills, Python proficiency, SQL, and experience in B2B SaaS and A/B testing platforms. | Eval GateAgent | 7 |