OpenAI logoOpenAI logo

Senior Researcher in Misalignment Research

OpenAISan Francisco
On-site Full-time

Clicking Apply Now takes you to AutoApply where you can tailor your resume and apply.


Experience Level

Senior

Qualifications

Applicants should possess a strong background in AI safety, misalignment research, or related fields. A proven track record in conducting rigorous evaluations, developing automated testing tools, and collaborating in multidisciplinary teams is essential. Candidates with experience in adversarial machine learning, system-level testing, and research publication are highly desirable.

About the job

About Our Team

At OpenAI, our Safety Systems team is dedicated to advancing the mission of developing and deploying safe artificial general intelligence (AGI). We are establishing a specialized research team focused on identifying and addressing critical misalignment issues that may arise as AGI technology evolves. Our goal is to proactively quantify and mitigate potential misalignment risks to ensure they do not threaten societal wellbeing.

Our research efforts are structured around four key areas:

  1. Worst-Case Demonstrations – Create compelling demonstrations that illustrate how AI systems can fail, particularly in scenarios where misaligned AGI could undermine human interests.

  2. Adversarial & Frontier Safety Evaluations – Develop rigorous evaluations based on these demonstrations to measure dangerous capabilities and remaining risks, focusing on issues like deceptive behavior and power-seeking tendencies.

  3. System-Level Stress Testing – Construct automated infrastructure to stress-test entire product stacks, evaluating their robustness under extreme conditions and evolving the tests as systems improve.

  4. Alignment Stress-Testing Research – Analyze failures in mitigations and publish insights to inform strategy and develop next-generation safeguards, collaborating with other research teams for collective advancement.

About the Role

We are looking for a passionate Senior Researcher focused on AI safety and red-teaming. In this role, you will design and execute innovative attacks, contribute to adversarial evaluations, and deepen our understanding of how safety measures can fail, and how they can be improved. Your findings will significantly impact OpenAI's product releases and long-term safety strategies.

Key Responsibilities

  • Create and implement worst-case demonstrations that clarify AGI alignment risks for stakeholders, particularly in critical use cases.

  • Develop comprehensive adversarial and system-level evaluations based on these demonstrations, promoting their integration across OpenAI.

  • Design automated tools and frameworks to enhance our red-teaming and stress-testing capabilities.

About OpenAI

OpenAI is at the forefront of AI innovation, dedicated to ensuring that artificial general intelligence is developed safely and responsibly. Our mission is to ensure that AGI benefits all of humanity, and our Safety Systems team plays a critical role in achieving this goal.

Similar jobs

Browse all companies, explore by city & role, or SEO search pages.

Tailoring 0 resumes

We'll move completed jobs to Ready to Apply automatically.