Guardrails
115 AI roles tagged guardrails.
Sector
Status
FilteredsectorAI Frontier×
| Company | Title | Sector | AI score | Other tags |
|---|---|---|---|---|
| OpenAI | Researcher, Misalignment Research | AI Frontier | 10 | Evals · Agent research · Frontier research |
| OpenAI | Researcher, Loss of Control | AI Frontier | 10 | Agent orchestration · Tool use · Evals · LLM observability · Agent research |
| Anthropic | Research Engineer, Frontier Red Team (Autonomy) | AI Frontier | 10 | Agent orchestration · Tool use · Evals · Embodied AI · RL robotics · Agent research |
| Anthropic | Anthropic AI Safety Fellow, UK | AI Frontier | 10 | Frontier research · Interpretability · Evals · RLHF |
| Anthropic | Anthropic AI Safety Fellow, US | AI Frontier | 10 | Frontier research · Interpretability · Evals · RL post-training |
| Mistral AI | Model Behavior Architect | AI Frontier | 9 | Evals · LLM observability · Agent orchestration · Tool use · Fine-tuning · RL post-training |
| OpenAI | Researcher, Alignment Science | AI Frontier | 9 | RL post-training · Evals · LLM observability · Interpretability |
| Anthropic | Research Engineer, Safeguards Labs | AI Frontier | 9 | Evals · Agent orchestration · Agent research · Fine-tuning · RL post-training |
| Writer | Security engineer, detection and response (US) | AI Frontier | 9 | Agent orchestration · LLM observability · Inference infra · Model serving |
| Writer | Security engineer, detection and response (UK) | AI Frontier | 9 | Model serving · Inference infra · LLM observability |
| Anthropic | Anthropic Fellows Program — AI Safety | AI Frontier | 9 | Interpretability · Evals · RL post-training |
| OpenAI | Researcher, Safety & Privacy | AI Frontier | 9 | Interpretability · Evals · LLM observability |
| Perplexity | Member of Technical Staff (Secure Intelligence Institute) | AI Frontier | 9 | Agent orchestration · Evals · Agent research |
| OpenAI | Machine Learning Engineer, Integrity | AI Frontier | 9 | Fine-tuning · Model serving · LLM observability · Evals |
| Anthropic | Security Labs Engineer | AI Frontier | 9 | Inference infra · Model serving |
| Anthropic | Research Lead, Training Insights | AI Frontier | 9 | Evals · LLM observability · Agent research · RL post-training · Frontier research |
| OpenAI | Threat Modeler, Preparedness | AI Frontier | 9 | Evals · Interpretability · Agent research |
| OpenAI | Researcher, Automated Red Teaming | AI Frontier | 9 | Evals · Agent orchestration · Tool use · LLM observability |
| OpenAI | Researcher, Frontier Cybersecurity Risks | AI Frontier | 9 | Agent orchestration · LLM observability · Evals · Model serving |
| Anthropic | Prompt Engineer, Agent Prompts & Evals | AI Frontier | 9 | Agent orchestration · Evals · LLM observability · Fine-tuning · Model serving |
| Anthropic | Research Scientist, Frontier Red Team (Emerging Risks) | AI Frontier | 9 | Evals · Agent research · LLM observability · Embodied AI |
| Anthropic | Research Engineer / Scientist, Frontier Red Team (Cyber) | AI Frontier | 9 | Agent orchestration · Evals · LLM observability |
| Anthropic | Applied Safety Research Engineer, Safeguards | AI Frontier | 9 | Evals · LLM observability · Fine-tuning · Agent research |
| Anthropic | Anthropic Fellows Program — AI Security | AI Frontier | 9 | Evals · LLM observability · Agent research · Frontier research |
| Anthropic | Cross-functional Prompt Engineer | AI Frontier | 9 | Fine-tuning · Evals · Agent orchestration · LLM observability |
| Anthropic | Research Engineer, Model Evaluations | AI Frontier | 9 | Evals · LLM observability · Fine-tuning · Model serving · Agent research · RL post-training |
| OpenAI | Offensive Security Engineer, Agent Products | AI Frontier | 9 | Agent orchestration · Tool use · Evals · Model serving · Inference infra |
| Anthropic | ML/Research Engineer, Safeguards | AI Frontier | 9 | Agent orchestration · Evals · Synthetic data · Agent research |
| Character AI | Research Engineer, AI Safety & Alignment | AI Frontier | 9 | Evals · Interpretability · RL post-training · Fine-tuning · LLM observability |
| OpenAI | Technical Lead, Safety Research | AI Frontier | 9 | RL post-training · Evals · Frontier research · Interpretability |