Anthropic

About Anthropic

Building safe and reliable AI systems for everyone

🏢 Tech👥 1001+ employees📅 Founded 2021📍 SoMa, San Francisco, CA💰 $29.3b4.5
B2BArtificial IntelligenceDeep TechMachine LearningSaaS

Key Highlights

  • Headquartered in SoMa, San Francisco, CA
  • Raised $29.3 billion in funding, including $13 billion Series F
  • Over 1,000 employees focused on AI safety and research
  • Launched Claude, an AI chat assistant rivaling ChatGPT

Anthropic, headquartered in SoMa, San Francisco, is an AI safety and research company focused on developing reliable, interpretable, and steerable AI systems. With over 1,000 employees and backed by Google, Anthropic has raised $29.3 billion in funding, including a monumental Series F round of $13 b...

🎁 Benefits

Anthropic offers comprehensive health, dental, and vision insurance for employees and their dependents, along with inclusive fertility benefits via Ca...

🌟 Culture

Anthropic's culture is rooted in AI safety and reliability, with a focus on producing less harmful outputs compared to existing AI systems. The compan...

Anthropic

Safeguards Analyst Mid-Level

AnthropicSan Francisco

Posted 9h agoMid-LevelSafeguards Analyst📍 San Francisco📍 New York💰 $230,000 - $310,000 / yearly
Apply Now →

Overview

Anthropic is hiring a Safeguards Analyst to focus on Account Abuse. You'll develop frameworks for detecting and preventing account abuse on the platform. This role requires a strong understanding of data analysis and operational capabilities.

Job Description

Who you are

You have a strong analytical mindset and are passionate about ensuring the safety and integrity of AI systems. You understand the complexities of account abuse and are committed to developing effective prevention frameworks. You thrive in environments where you can engage with diverse data sources and collaborate with cross-functional teams to enhance detection capabilities.

You are comfortable working with explicit content and understand the implications of engaging with sensitive topics. Your experience allows you to evaluate and integrate third-party vendor signals effectively, ensuring that new data sources genuinely improve detection coverage. You are proactive in expanding internal account signals with new behavioral indicators, demonstrating your commitment to continuous improvement.

What you'll do

In this role, you will develop and iterate on account signals and prevention frameworks that consolidate internal and external data into actionable abuse indicators. You will optimize identity and account-linking signals using graph-based data infrastructure to detect coordinated and scaled account abuse. Your responsibilities will include evaluating, integrating, and operationalizing third-party vendor signals, assessing whether new data sources provide genuine lift in detection. You will also expand internal account signals with new data sources and behavioral indicators to improve detection coverage.

What we offer

Anthropic offers competitive compensation and benefits, including optional equity donation matching, generous vacation and parental leave, and flexible working hours. You will work in a collaborative office space in San Francisco, contributing to a mission-driven organization focused on building beneficial AI systems.

Interested in this role?

Apply now or save it for later. Get alerts for similar jobs at Anthropic.

Similar Jobs You Might Like

Based on your interests and this role

Anthropic

Ai Research Engineer

Anthropic📍 San Francisco - On-Site

Anthropic is hiring a Privacy Research Engineer to design and implement privacy-preserving techniques for AI systems. You'll work with Python and ML frameworks like PyTorch and JAX in San Francisco. This position requires experience in privacy-preserving machine learning.

🏛️ On-SiteMid-Level
9h ago
Anthropic

Software Engineering

Anthropic📍 San Francisco

Anthropic is seeking Software Engineers for their Safeguards team to develop safety mechanisms for AI systems. You'll work with Java and Python to build monitoring systems and abuse detection infrastructure. This role requires 5-10 years of experience in software engineering.

Mid-Level
9h ago
Anthropic

Machine Learning Engineer

Anthropic📍 San Francisco

Anthropic is hiring ML/Research Engineers to develop systems that detect and mitigate misuse of AI technologies. You'll work with Python and machine learning frameworks like TensorFlow and PyTorch. This role requires experience in building classifiers and monitoring systems for AI safety.

Mid-Level
9h ago
Anthropic

Applied Scientist

Anthropic📍 San Francisco

Anthropic is seeking an Applied Safety Research Engineer to develop methods for evaluating AI safety. You'll work with machine learning and Python to design experiments that improve model evaluations. This role requires a research-oriented mindset and experience in applied ML.

Mid-Level
9h ago
Anthropic

Other Technical Roles

Anthropic📍 San Francisco

Anthropic is seeking a Technical Scaled Abuse Threat Investigator to join their Threat Intelligence team. You'll be responsible for detecting and investigating large-scale misuse of AI systems. This role requires strong analytical skills and experience in threat detection.

Mid-Level
9h ago