
About Anthropic
Building safe and reliable AI systems for everyone
Key Highlights
- Headquartered in SoMa, San Francisco, CA
- Raised $29.3 billion in funding, including $13 billion Series F
- Over 1,000 employees focused on AI safety and research
- Launched Claude, an AI chat assistant rivaling ChatGPT
Anthropic, headquartered in SoMa, San Francisco, is an AI safety and research company focused on developing reliable, interpretable, and steerable AI systems. With over 1,000 employees and backed by Google, Anthropic has raised $29.3 billion in funding, including a monumental Series F round of $13 b...
🎁 Benefits
Anthropic offers comprehensive health, dental, and vision insurance for employees and their dependents, along with inclusive fertility benefits via Ca...
🌟 Culture
Anthropic's culture is rooted in AI safety and reliability, with a focus on producing less harmful outputs compared to existing AI systems. The compan...
Overview
Anthropic is hiring a Safeguards Analyst to focus on Account Abuse. You'll develop frameworks for detecting and preventing account abuse on the platform. This role requires a strong understanding of data analysis and operational capabilities.
Job Description
Who you are
You have a strong analytical mindset and are passionate about ensuring the safety and integrity of AI systems. You understand the complexities of account abuse and are committed to developing effective prevention frameworks. You thrive in environments where you can engage with diverse data sources and collaborate with cross-functional teams to enhance detection capabilities.
You are comfortable working with explicit content and understand the implications of engaging with sensitive topics. Your experience allows you to evaluate and integrate third-party vendor signals effectively, ensuring that new data sources genuinely improve detection coverage. You are proactive in expanding internal account signals with new behavioral indicators, demonstrating your commitment to continuous improvement.
What you'll do
In this role, you will develop and iterate on account signals and prevention frameworks that consolidate internal and external data into actionable abuse indicators. You will optimize identity and account-linking signals using graph-based data infrastructure to detect coordinated and scaled account abuse. Your responsibilities will include evaluating, integrating, and operationalizing third-party vendor signals, assessing whether new data sources provide genuine lift in detection. You will also expand internal account signals with new data sources and behavioral indicators to improve detection coverage.
What we offer
Anthropic offers competitive compensation and benefits, including optional equity donation matching, generous vacation and parental leave, and flexible working hours. You will work in a collaborative office space in San Francisco, contributing to a mission-driven organization focused on building beneficial AI systems.
Interested in this role?
Apply now or save it for later. Get alerts for similar jobs at Anthropic.
Similar Jobs You Might Like
Based on your interests and this role

Ai Research Engineer
Anthropic is hiring a Privacy Research Engineer to design and implement privacy-preserving techniques for AI systems. You'll work with Python and ML frameworks like PyTorch and JAX in San Francisco. This position requires experience in privacy-preserving machine learning.

Software Engineering
Anthropic is seeking Software Engineers for their Safeguards team to develop safety mechanisms for AI systems. You'll work with Java and Python to build monitoring systems and abuse detection infrastructure. This role requires 5-10 years of experience in software engineering.

Machine Learning Engineer
Anthropic is hiring ML/Research Engineers to develop systems that detect and mitigate misuse of AI technologies. You'll work with Python and machine learning frameworks like TensorFlow and PyTorch. This role requires experience in building classifiers and monitoring systems for AI safety.

Applied Scientist
Anthropic is seeking an Applied Safety Research Engineer to develop methods for evaluating AI safety. You'll work with machine learning and Python to design experiments that improve model evaluations. This role requires a research-oriented mindset and experience in applied ML.

Other Technical Roles
Anthropic is seeking a Technical Scaled Abuse Threat Investigator to join their Threat Intelligence team. You'll be responsible for detecting and investigating large-scale misuse of AI systems. This role requires strong analytical skills and experience in threat detection.