This list contains only the countries for which job offers have been published in the selected language (e.g., in the French version, only job offers written in French are displayed, and in the English version, only those in English).
The Alignment team at OpenAI is dedicated to ensuring that our AI systems are safe, trustworthy, and consistently aligned with human values, even as they scale in complexity and capability. Our work is at the cutting edge of AI research, focusing on developing methodologies that enable AI to robustly follow human intent across a wide range of scenarios, including those that are adversarial or high-stakes. We concentrate on the most pressing challenges, ensuring our work addresses areas where AI could have the most significant consequences. By focusing on risks that we can quantify and where our efforts can make a tangible difference, we aim to ensure that our models are ready for the complex, real-world environments in which they will be deployed. The two pillars of our approach are: (1) harnessing improved capabilities into alignment, making sure that our alignment techniques improve, rather than break, as capabilities grow, and (2) centering humans by developing mechanisms and interfaces that enable humans to both express their intent and to effectively supervise and control AIs, even in highly complex situations.
Job Responsibility:
Develop and evaluate alignment capabilities that are subjective, context-dependent, and hard to measure
Design evaluations to reliably measure risks and alignment with human intent and values
Build tools and evaluations to study and test model robustness in different situations
Design experiments to understand laws for how alignment scales as a function of compute, data, lengths of context and action, as well as resources of adversaries
Design and evaluate new Human-AI-interaction paradigms and scalable oversight methods that redefine how humans interact with, understand, and supervise our models
Train model to be calibrated on correctness and risk
Designing novel approaches for using AI in alignment research
Requirements:
PhD or equivalent experience in research in computer science, computational science, data science, cognitive science, or similar fields
Strong engineering skills, particularly in designing and optimizing large-scale machine learning systems (e.g., PyTorch)
Deep understanding of the science behind alignment algorithms and techniques
Can develop data visualization or data collection interfaces (e.g., TypeScript, Python)
Enjoy fast-paced, collaborative, and cutting-edge research environments
Want to focus on developing AI models that are trustworthy, safe, and reliable, especially in high-stakes scenarios
Nice to have:
Are a team player – willing to do a variety of tasks that move the team forward
What we offer:
Medical, dental, and vision insurance for you and your family, with employer contributions to Health Savings Accounts
Pre-tax accounts for Health FSA, Dependent Care FSA, and commuter expenses (parking and transit)
401(k) retirement plan with employer match
Paid parental leave (up to 24 weeks for birth parents and 20 weeks for non-birthing parents), plus paid medical and caregiver leave (up to 8 weeks)
Paid time off: flexible PTO for exempt employees and up to 15 days annually for non-exempt employees
13+ paid company holidays, and multiple paid coordinated company office closures throughout the year for focus and recharge, plus paid sick or safe time (1 hour per 30 hours worked, or more, as required by applicable state or local law)
Mental health and wellness support
Employer-paid basic life and disability coverage
Annual learning and development stipend to fuel your professional growth
Daily meals in our offices, and meal delivery credits as eligible
Relocation support for eligible employees
Additional taxable fringe benefits, such as charitable donation matching and wellness stipends, may also be provided
Offers Equity
Performance-related bonus(es) for eligible employees