Demo

Research Engineer / Scientist, Safety Reasoning

The Rundown AI, Inc.
San Francisco, CA Full Time
POSTED ON 3/1/2025
AVAILABLE BEFORE 5/18/2025

About the TeamThe Safety Systems team is responsible for various safety work to ensure our best models can be safely deployed to the real world to benefit society and is at the forefront of OpenAI's mission to build and deploy safe AGI, driving our commitment to AI safety and fostering a culture of trust and transparency.The Safety Reasoning Research team is poised at the intersection of short-term pragmatic projects and long-term fundamental research, prioritizing rapid system development while maintaining technical robustness. Key focus areas include improving foundational models’ ability to accurately reason about safety, values, and questions of cultural norms, refining moderation models, driving rapid policy improvements, and addressing critical societal challenges like election misinformation. As we venture into 2024, the team seeks talents adept in novel abuse discovery and policy iteration, aligning with our high-priority goals of multimodal moderation and ensuring digital safety.About the RoleThe role involves developing innovative machine learning techniques that push the limit of our foundation model’s safety understanding and capability. You will engage in defining and developing realistic and impactful safety tasks that, once improved, can be integrated into OpenAI's safety systems or benefit other safety / alignment research initiatives. Examples of safety initiatives include moderation policy enforcement, policy development using democratic input, and safety reward modeling. You will be experimenting with a wide range of research techniques not limited to reasoning, architecture, data, and multimodal.In this role, you will : Conduct applied research to improve the ability of foundational models to accurately reason about questions of human values, morals, ethics, and cultural norms, and apply these improved models to practical safety challenges.Develop and refine AI moderation models to detect and mitigate known and emerging patterns of AI misuse and abuse.Work with policy researchers to adapt and iterate on our content policies to ensure effective prevention of harmful behavior.Contribute to research on multimodal content analysis to enhance our moderation capabilities.Develop and improve pipelines for automated data labeling and augmentation, model training, evaluation, and deployment, including active learning processes, routines for calibration, and validation data refresh.Experiment and design an effective red-teaming pipeline to examine the robustness of our harm prevention systems and identify areas for future improvement.You might thrive in this role if you : Are excited about OpenAI’s mission of building safe, universally beneficial AGI and are aligned with OpenAI’s charter .Possess 5 years of research engineering experience and proficiency in Python or similar languages.Thrive in environments involving large-scale AI systems and multimodal datasets (a plus).Exhibit proficiency in the field of AI safety, focusing on topics like RLHF, adversarial training, robustness, fairness & biases, which is extremely advantageous.Show enthusiasm for AI safety and dedication to enhancing the safety of cutting-edge AI models for real-world use.About OpenAIOpenAI is an AI research and deployment company dedicated to ensuring that general-purpose artificial intelligence benefits all of humanity. We push the boundaries of the capabilities of AI systems and seek to safely deploy them to the world through our products. AI is an extremely powerful tool that must be created with safety and human needs at its core, and to achieve our mission, we must encompass and value the many different perspectives, voices, and experiences that form the full spectrum of humanity.We are an equal opportunity employer and do not discriminate on the basis of race, religion, national origin, gender, sexual orientation, age, veteran status, disability or any other legally protected status.For US Based Candidates : Pursuant to the San Francisco Fair Chance Ordinance, we will consider qualified applicants with arrest and conviction records.We are committed to providing reasonable accommodations to applicants with disabilities, and requests can be made via this link .At OpenAI, we believe artificial intelligence has the potential to help people solve immense global challenges, and we want the upside of AI to be widely shared. Join us in shaping the future of technology.#J-18808-Ljbffr

If your compensation planning software is too rigid to deploy winning incentive strategies, it’s time to find an adaptable solution. Compensation Planning
Enhance your organization's compensation strategy with salary data sets that HR and team managers can use to pay your staff right. Surveys & Data Sets

What is the career path for a Research Engineer / Scientist, Safety Reasoning?

Sign up to receive alerts about other jobs on the Research Engineer / Scientist, Safety Reasoning career path by checking the boxes next to the positions that interest you.
Income Estimation: 
$113,077 - $147,784
Income Estimation: 
$135,356 - $164,911
Income Estimation: 
$153,902 - $198,246
Income Estimation: 
$113,077 - $147,784
Income Estimation: 
$135,356 - $164,911
Income Estimation: 
$153,902 - $198,246
Income Estimation: 
$135,356 - $164,911
Income Estimation: 
$153,053 - $187,211
Income Estimation: 
$153,902 - $198,246
Income Estimation: 
$98,763 - $126,233
Income Estimation: 
$116,330 - $143,011
Income Estimation: 
$113,077 - $147,784
Income Estimation: 
$116,330 - $143,011
Income Estimation: 
$135,356 - $164,911
Income Estimation: 
$153,902 - $198,246
View Core, Job Family, and Industry Job Skills and Competency Data for more than 15,000 Job Titles Skills Library

Job openings at The Rundown AI, Inc.

The Rundown AI, Inc.
Hired Organization Address San Francisco, CA Full Time
About the team OpenAI’s People team is dedicated to hiring, engaging, and retaining world-class talent to safely develop...
The Rundown AI, Inc.
Hired Organization Address San Francisco, CA Full Time
Meta seeks an attorney with experience in online intellectual property issues to serve as Associate General Counsel, IP ...
The Rundown AI, Inc.
Hired Organization Address Washington, DC Full Time
About Shield AI Founded in 2015, Shield AI is a venture-backed defense technology company focused on protecting service ...
The Rundown AI, Inc.
Hired Organization Address New York, NY Full Time
Life at UiPath The people at UiPath believe in the transformative power of automation to change how the world works. We’...

Not the job you're looking for? Here are some other Research Engineer / Scientist, Safety Reasoning jobs in the San Francisco, CA area that may be a better fit.

Research Engineer - Reasoning

Prime Intellect, Inc., San Francisco, CA

AI Assistant is available now!

Feel free to start your new journey!