
About Anthropic
Building safe and reliable AI systems for everyone
Key Highlights
- Headquartered in SoMa, San Francisco, CA
- Raised $29.3 billion in funding, including $13 billion Series F
- Over 1,000 employees focused on AI safety and research
- Launched Claude, an AI chat assistant rivaling ChatGPT
Anthropic, headquartered in SoMa, San Francisco, is an AI safety and research company focused on developing reliable, interpretable, and steerable AI systems. With over 1,000 employees and backed by Google, Anthropic has raised $29.3 billion in funding, including a monumental Series F round of $13 b...
🎁 Benefits
Anthropic offers comprehensive health, dental, and vision insurance for employees and their dependents, along with inclusive fertility benefits via Ca...
🌟 Culture
Anthropic's culture is rooted in AI safety and reliability, with a focus on producing less harmful outputs compared to existing AI systems. The compan...
Overview
Anthropic is hiring an AI Research Engineer to work on mechanistic interpretability of AI systems. You'll collaborate with a team focused on understanding how neural networks function. This role requires a strong background in AI and machine learning.
Job Description
Who you are
You are a researcher or engineer with a keen interest in AI interpretability — you wonder how modern language models work and how we can trust them. You have a background in AI or machine learning and are excited about the challenge of reverse-engineering neural networks to understand their mechanisms. You are committed to building safe and beneficial AI systems and are eager to contribute to a team that values collaboration and innovation.
You understand that interpretability can mean many things, but you are particularly focused on mechanistic interpretability — you aim to discover how neural network parameters map to meaningful algorithms. You are familiar with the challenges of building tools that help visualize and analyze the inner workings of AI models, akin to using microscopes in biology or neuroscience.
What you'll do
Join Anthropic's Interpretability team and contribute to groundbreaking research aimed at making AI systems more interpretable and reliable. You will work closely with a diverse group of researchers and engineers to develop methodologies for understanding neural networks. Your role will involve designing experiments, analyzing data, and collaborating with team members to publish findings that advance the field of AI interpretability.
You will engage in discussions about the implications of your work on AI safety and ethics, ensuring that the systems we build are not only effective but also trustworthy. You will have the opportunity to contribute to publications and share insights with the broader AI community, helping to shape the future of AI technology.
What we offer
Anthropic provides a supportive work environment where you can thrive as part of a mission-driven team. We offer competitive compensation and benefits, including generous vacation and parental leave, flexible working hours, and a collaborative office space in San Francisco. You will have the chance to work on meaningful projects that aim to create beneficial AI systems for society. We encourage you to apply even if your experience doesn't match every requirement, as we value diverse perspectives and backgrounds.
Interested in this role?
Apply now or save it for later. Get alerts for similar jobs at Anthropic.
Similar Jobs You Might Like
Based on your interests and this role

Ai Research Engineer
OpenAI is hiring a Researcher in Interpretability to develop and publish research on understanding deep learning models. You'll work with Python and machine learning techniques in San Francisco. This role requires a strong background in AI safety and mechanistic interpretability.

Research Scientist
Anthropic is hiring a Research Scientist focused on Interpretability to explore how modern language models work and ensure their safety. You'll engage in mechanistic interpretability research to enhance AI systems. This role requires a strong background in AI and machine learning.

Research Scientist
Anthropic is seeking a Research Manager focused on interpretability to contribute to AI safety research. You'll work with a team dedicated to understanding neural networks and their mechanisms. This role is ideal for those passionate about AI interpretability.

Ai Engineer
Factory is hiring a Research Engineer to design and integrate advanced AI and ML capabilities. You'll work on AI-driven systems and collaborate with product and engineering teams. This position requires 2+ years of experience in AI/ML engineering.

Ai Research Engineer
OpenAI is hiring a Research Engineer to build AI systems that achieve unprecedented levels of performance. You'll work with technologies like Python and deep learning frameworks in San Francisco. This role requires strong programming skills and experience with large distributed systems.