companyAnthropic logo

Research Engineer / Scientist, Alignment Science

AnthropicSan Francisco, CA
On-site Full-time

Clicking Apply Now takes you to AutoApply where you can tailor your resume and apply.


Unlock Your Potential

Generate Job-Optimized Resume

One Click And Our AI Optimizes Your Resume to Match The Job Description.

Is Your Resume Optimized For This Role?

Find Out If You're Highlighting The Right Skills And Fix What's Missing

Qualifications

Candidates should possess a strong background in machine learning, statistics, or a related field, along with hands-on experience in developing and executing complex experiments. Proficiency in programming languages such as Python and familiarity with machine learning frameworks is essential. A keen interest in AI safety and ethical implications is highly valued. Prior experience in collaborative research environments and excellent problem-solving skills will be beneficial.

About the job

About Anthropic

At Anthropic, we are driven by our mission to develop reliable, interpretable, and steerable AI systems. Our commitment is to ensure that AI is safe and beneficial not only for our users but also for society as a whole. Our rapidly expanding team comprises dedicated researchers, engineers, policy specialists, and business leaders collaborating to create impactful AI technologies.

About the Role:

As a Research Engineer focusing on Alignment Science, you will design and execute sophisticated machine learning experiments aimed at understanding and guiding the behavior of advanced AI systems. Your passion lies in making AI systems helpful, honest, and safe, particularly in the face of challenges posed by human-level capabilities. You embody both the scientific and engineering mindsets. In this role, you will engage in exploratory research on AI safety, concentrating on risks associated with future powerful systems (such as those classified as ASL-3 or ASL-4 under our Responsible Scaling Policy), often working in collaboration with teams focused on Interpretability, Fine-Tuning, and the Frontier Red Team.
 
Discover more about our current research topics and insights on our blog, as we delve into pressing issues such as:
  • Scalable Oversight: Innovating techniques to ensure that highly capable models remain helpful and truthful, even as they exceed human-level intelligence.
  • AI Control: Developing strategies to maintain the safety and harmlessness of advanced AI systems in novel or adversarial environments.
  • Alignment Stress Testing: Implementing rigorous testing frameworks to evaluate AI alignment under various conditions.

About Anthropic

Anthropic is at the forefront of AI innovation, dedicated to creating AI systems that are not only effective but also responsible and ethical. Our expert team collaborates across disciplines to tackle the most pressing challenges in AI development, ensuring our technologies benefit humanity. Join us in our mission to shape the future of AI!

Similar jobs

Tailoring 0 resumes

We'll move completed jobs to Ready to Apply automatically.