Anthropic
17 open roles (AI safety, policy & security)
description
Anthropic is a leading AI research and product company that aims to build reliable, interpretable, and steerable AI systems. Their research interests span multiple areas including natural language, human feedback, scaling laws, reinforcement learning, code generation, and interpretability. Recently announced teams working on safety issues include their Frontier Red Team (working on adversarial testing of advanced ML models), and their Alignment Stress-Testing Team (working on red-teaming Anthropic's alignment and evaluation efforts).
In our career review on working at a leading AI lab you can find some discussion on various considerations for and against this option.
Open roles (AI safety, policy & security )Some roles at Anthropic are focused on advancing AI capabilities (which helps enable some of their safety research especially around large model evaluations and interpretability).
For some discussion of working on AI capabilities to advance AI safety work see some anonymous advice we've collected from experts on this topic.
You can find all of Anthropic's roles on their careers page.
Some roles at Anthropic are focused on advancing AI capabilities (which helps enable some of their safety research especially around large model evaluations and interpretability).
For some discussion of working on AI capabilities to advance AI safety work see some anonymous advice we've collected from experts on this topic.
You can find all of Anthropic's roles on their careers page.
Learn more
80,000 Hours links
Preventing an AI-related catastrophe
Problem profile
Working at a leading AI lab
Career review
AI safety technical research
Career review
Anonymous advice on if you should work on AI capabilities to help reduce AI risk
Article
Interview with Chris Olah, who leads Anthropic's research into interpretability
Podcast
Interview with Nova DasSarma, who works on information security at Anthropic
Podcast
External content
Anthropic's core views on AI safety
Information
An interview with Anthropic's co-founders Daniela and Dario Amodei's (Future of Life Institute podcast)
Podcast
Anthropic's AI safety research papers
Research