This list contains only the countries for which job offers have been published in the selected language (e.g., in the French version, only job offers written in French are displayed, and in the English version, only those in English).
As a Research Scientist focused on AI Controls and Monitoring, you will design methods, systems, and experiments to ensure that advanced AI models and agents remain aligned with intended goals, even in high-stakes or adversarial environments.
Job Responsibility:
Design methods, systems, and experiments to ensure advanced AI models and agents remain aligned with intended goals
Develop monitoring techniques and observability methods to track AI behavior in real time
Research mechanisms for layered control, including fail-safes, oversight protocols, and intervention methods
Design red-team simulations to probe weaknesses in oversight and control mechanisms
Build mitigations to close identified gaps
Collaborate with policymakers, engineers, and other researchers to establish standards and benchmarks
Requirements:
Commitment to mission of promoting safe, secure, and trustworthy AI deployments
Practical experience conducting technical research collaboratively
Comfort designing control and monitoring experiments for AI systems
Experience building prototype systems
Ability to turn research ideas into working prototypes
Track record of published research in machine learning, particularly generative AI
At least three years of experience addressing sophisticated ML problems
Strong written and verbal communication skills
Nice to have:
Experience with runtime monitoring, anomaly detection, or observability for ML systems
Familiarity with AI control or alignment research (e.g., scalable oversight, interpretability, debate)
Experience with post-training and RL techniques such as RLHF, DPO, GRPO