Demo

Research Engineer / Research Scientist, Alignment

OpenAI
OpenAI Salary
San Francisco, CA Full Time
POSTED ON 1/14/2025
AVAILABLE BEFORE 4/8/2025

About the Team

The Alignment team at OpenAI is dedicated to ensuring that our AI systems are safe, trustworthy, and consistently aligned with human values, even as they scale in complexity and capability. Our work is at the cutting edge of AI research, focusing on developing methodologies that enable AI to robustly follow human intent across a wide range of scenarios, including those that are adversarial or high-stakes. We concentrate on the most pressing challenges, ensuring our work addresses areas where AI could have the most significant consequences. By focusing on risks that we can quantify and where our efforts can make a tangible difference, we aim to ensure that our models are ready for the complex, real-world environments in which they will be deployed.

The two pillars of our approach are : harnessing improved capabilities into alignment, making sure that our alignment techniques improve, rather than break, as capabilities grow, and centering humans by developing mechanisms and interfaces that enable humans to both express their intent and to effectively supervise and control AIs, even in highly complex situations.

About the Role

As a Research Engineer / Research Scientist on the Alignment team, you will be at the forefront of ensuring that our AI systems consistently follow human intent, even in complex and unpredictable scenarios. Your role will involve designing and implementing scalable solutions that ensure the alignment of AI as their capabilities grow and that integrate human oversight into AI decision-making.

This role is based in San Francisco, CA. We use a hybrid work model of 3 days in the office per week and offer relocation assistance to new employees.

In this role, you will :

We are seeking research engineers and research scientists to help design and implement experiments for alignment research. Responsibilities may include :

Develop and evaluate alignment capabilities that are subjective, context-dependent, and hard to measure.

Design evaluations to reliably measure risks and alignment with human intent and values.

Build tools and evaluations to study and test model robustness in different situations.

Design experiments to understand laws for how alignment scales as a function of compute, data, lengths of context and action, as well as resources of adversaries.

Design and evaluate new Human-AI-interaction paradigms and scalable oversight methods that redefine how humans interact with, understand, and supervise our models.

Train model to be calibrated on correctness and risk.

Designing novel approaches for using AI in alignment research

You might thrive in this role if you :

Are a team player – willing to do a variety of tasks that move the team forward.

Have a PhD or equivalent experience in research in computer science, computational science, data science, cognitive science, or similar fields.

Have strong engineering skills, particularly in designing and optimizing large-scale machine learning systems(e.g., PyTorch).

Have a deep understanding of the science behind alignment algorithms and techniques.

Can develop data visualization or data collection interfaces (e.g., TypeScript, Python).

Enjoy fast-paced, collaborative, and cutting-edge research environments.

Want to focus on developing AI models that are trustworthy, safe, and reliable, especially in high-stakes scenarios.

If your compensation planning software is too rigid to deploy winning incentive strategies, it’s time to find an adaptable solution. Compensation Planning
Enhance your organization's compensation strategy with salary data sets that HR and team managers can use to pay your staff right. Surveys & Data Sets

What is the career path for a Research Engineer / Research Scientist, Alignment?

Sign up to receive alerts about other jobs on the Research Engineer / Research Scientist, Alignment career path by checking the boxes next to the positions that interest you.
Income Estimation: 
$113,077 - $147,784
Income Estimation: 
$135,356 - $164,911
Income Estimation: 
$153,902 - $198,246
Income Estimation: 
$82,813 - $108,410
Income Estimation: 
$120,989 - $162,093
Income Estimation: 
$74,806 - $91,633
Income Estimation: 
$71,928 - $87,026
Income Estimation: 
$145,337 - $174,569
Income Estimation: 
$98,763 - $126,233
Income Estimation: 
$116,330 - $143,011
Income Estimation: 
$113,077 - $147,784
Income Estimation: 
$116,330 - $143,011
Income Estimation: 
$135,356 - $164,911
Income Estimation: 
$153,902 - $198,246
Income Estimation: 
$102,775 - $137,396
Income Estimation: 
$153,127 - $203,425
Income Estimation: 
$139,626 - $193,276
Income Estimation: 
$164,650 - $211,440
Income Estimation: 
$130,030 - $173,363
View Core, Job Family, and Industry Job Skills and Competency Data for more than 15,000 Job Titles Skills Library

Job openings at OpenAI

OpenAI
Hired Organization Address San Francisco, CA Full Time
About the Team The Special Projects team is dedicated to tackling high-priority initiatives that span across strategic i...
OpenAI
Hired Organization Address San Francisco, CA Full Time
About the Team OpenAI’s mission is to ensure that general-purpose artificial intelligence benefits all of humanity. We b...
OpenAI
Hired Organization Address Washington, DC Full Time
About the Team The Corporate Security team is responsible for the security and protection of all OpenAI employees and ex...
OpenAI
Hired Organization Address Washington, DC Full Time
About The Team The Engagement Management team ensures the successful post-sales deployment and value realization of Chat...

Not the job you're looking for? Here are some other Research Engineer / Research Scientist, Alignment jobs in the San Francisco, CA area that may be a better fit.

Research Engineer, Alignment Science

Karkidi, San Francisco, CA

AI Assistant is available now!

Feel free to start your new journey!