
About Anthropic
Building safe and reliable AI systems for everyone
Key Highlights
- Headquartered in SoMa, San Francisco, CA
- Raised $29.3 billion in funding, including $13 billion Series F
- Over 1,000 employees focused on AI safety and research
- Launched Claude, an AI chat assistant rivaling ChatGPT
Anthropic, headquartered in SoMa, San Francisco, is an AI safety and research company focused on developing reliable, interpretable, and steerable AI systems. With over 1,000 employees and backed by Google, Anthropic has raised $29.3 billion in funding, including a monumental Series F round of $13 b...
π Benefits
Anthropic offers comprehensive health, dental, and vision insurance for employees and their dependents, along with inclusive fertility benefits via Ca...
π Culture
Anthropic's culture is rooted in AI safety and reliability, with a focus on producing less harmful outputs compared to existing AI systems. The compan...
Overview
Anthropic is seeking Software Engineers for their Safeguards team to develop safety mechanisms for AI systems. You'll work with Java and Python to build monitoring systems and abuse detection infrastructure. This role requires 5-10 years of experience in software engineering.
Job Description
Who you are
You have a Bachelorβs degree in Computer Science, Software Engineering, or comparable experience, along with 5-10+ years of experience in a software engineering position, preferably with a focus on safety mechanisms in AI systems. You are skilled in programming languages such as Java and Python, and you have a strong understanding of building robust systems that can monitor and enforce safety protocols effectively. You are detail-oriented and have experience in developing monitoring systems that can detect unwanted behaviors from API partners. You thrive in collaborative environments and are eager to work with researchers and analysts to improve AI safety.
Desirable
Experience with machine learning frameworks and familiarity with AI safety principles would be a plus. You are comfortable analyzing user reports and have a proactive approach to identifying and mitigating risks associated with AI usage. You are passionate about building systems that prioritize user well-being and uphold ethical standards in technology.
What you'll do
As a Software Engineer on the Safeguards team, you will be responsible for developing monitoring systems that detect unwanted behaviors from our API partners and potentially take automated enforcement actions. You will surface these behaviors in internal dashboards for manual review by analysts. Your role will also involve building abuse detection mechanisms and infrastructure to surface abuse patterns to our research teams, helping to harden models at the training stage. You will work on creating robust and reliable multi-layered defenses for real-time improvement of safety mechanisms that work at scale. Additionally, you will analyze user reports of inappropriate content or accounts, ensuring that our AI systems operate within acceptable use policies.
What we offer
At Anthropic, we offer competitive compensation and benefits, including optional equity donation matching, generous vacation and parental leave, and flexible working hours. You will have the opportunity to work in a lovely office space in San Francisco or New York City, collaborating with a diverse team of committed researchers, engineers, and policy experts. We encourage you to apply even if your experience doesn't match every requirement, as we value diverse perspectives and backgrounds in our mission to create safe and beneficial AI systems.
Interested in this role?
Apply now or save it for later. Get alerts for similar jobs at Anthropic.
Similar Jobs You Might Like
Based on your interests and this role

Software Engineering
Anthropic is hiring a Software Engineer for their Safeguards Infrastructure team to build foundational systems for AI safety and oversight. You'll work with Python to develop robust mechanisms for monitoring and preventing misuse of AI models. This role requires 4-10+ years of experience in software engineering.

Software Engineering
OpenAI is hiring a Software Engineer for their AI Safety team to build trust and safety capabilities in AI models. You'll work with Python and AI techniques to ensure responsible deployment of advanced models. This position requires experience in software engineering and a strong understanding of AI safety.

Machine Learning Engineer
Anthropic is seeking a Machine Learning Infrastructure Engineer to build and scale critical infrastructure for AI safety systems. You'll work with distributed systems and machine learning technologies to ensure reliable operations. This role requires experience in building scalable ML infrastructure.

Software Engineering
Twitch is hiring a Software Engineer for their Safety ML team to develop systems that protect users from abuse. You'll work with machine learning and Python to build reliable services. This position requires experience in software engineering and machine learning.

Software Engineering
Amazon is hiring a Software Engineer for the Safety ML team to develop systems that protect users from abuse on Twitch. You'll work with machine learning and AWS technologies. This position requires experience in software engineering and machine learning.