
About Anthropic
Building safe and reliable AI systems for everyone
Key Highlights
- Headquartered in SoMa, San Francisco, CA
- Raised $29.3 billion in funding, including $13 billion Series F
- Over 1,000 employees focused on AI safety and research
- Launched Claude, an AI chat assistant rivaling ChatGPT
Anthropic, headquartered in SoMa, San Francisco, is an AI safety and research company focused on developing reliable, interpretable, and steerable AI systems. With over 1,000 employees and backed by Google, Anthropic has raised $29.3 billion in funding, including a monumental Series F round of $13 b...
🎁 Benefits
Anthropic offers comprehensive health, dental, and vision insurance for employees and their dependents, along with inclusive fertility benefits via Ca...
🌟 Culture
Anthropic's culture is rooted in AI safety and reliability, with a focus on producing less harmful outputs compared to existing AI systems. The compan...
Skills & Technologies
Overview
Anthropic is seeking a Research Scientist/Engineer focused on honesty to develop techniques that minimize hallucinations and enhance truthfulness in language models. You'll work primarily with Python to create robust AI systems. This role is based in New York City or San Francisco.
Job Description
Who you are
You are a dedicated researcher or engineer with a strong background in AI and machine learning, particularly in developing techniques that enhance the reliability and truthfulness of language models. Your expertise in Python allows you to design and implement innovative solutions that address complex challenges in AI systems. You are passionate about creating AI that is safe and beneficial for users and society, and you thrive in collaborative environments where you can work with a diverse team of experts.
You have experience in data curation and understand the importance of accuracy in training data. Your ability to develop classifiers to detect hallucinations or miscalibrated claims demonstrates your commitment to high standards in AI development. You are comfortable navigating the complexities of AI ethics and are motivated to contribute to the responsible advancement of technology.
What you'll do
In this role, you will spearhead the development of techniques to minimize hallucinations in language models, ensuring that they maintain high standards of accuracy and honesty. You will design and implement novel data curation pipelines to identify, verify, and filter training data, enhancing the overall quality of the models. Your work will involve developing specialized classifiers that can detect potential hallucinations or misleading claims made by the models, contributing to the integrity of AI systems.
You will collaborate closely with other researchers and engineers on the Finetuning Alignment team, sharing insights and strategies to improve the performance of AI systems. Your contributions will be critical in shaping the future of AI at Anthropic, as you work to create systems that are interpretable and steerable, aligning with the company's mission to build beneficial AI.
What we offer
Anthropic provides a supportive work environment where you can grow your skills and make a meaningful impact in the field of AI. We offer competitive compensation and benefits, including flexible working hours and generous vacation and parental leave. Our offices in New York City and San Francisco are designed to foster collaboration and creativity, allowing you to work alongside talented colleagues who share your commitment to ethical AI development. We encourage you to apply even if your experience doesn't match every requirement, as we value diverse perspectives and backgrounds in our team.
Interested in this role?
Apply now or save it for later. Get alerts for similar jobs at Anthropic.
Similar Jobs You Might Like
Based on your interests and this role

Research Scientist
OpenAI is hiring a Researcher for their Trustworthy AI team to study societal impacts of AI models. You'll work on translating policy problems into measurable research. This role is based in San Francisco.

Ai Research Engineer
Anthropic is seeking a Research Scientist/Engineer for their Alignment Finetuning team to develop techniques for training language models aligned with human values. You'll work with Python to implement novel finetuning techniques and improve model behavior.

Ai Research Engineer
Anthropic is hiring a Staff Research Engineer for their Discovery Team to work on developing advanced AI systems. You'll focus on identifying and addressing key blockers in the path to scientific AGI, utilizing skills in Python, Docker, and machine learning. This role requires familiarity with distributed systems and performance optimization.

Ai Research Engineer
Anthropic is hiring a Research Engineer / Scientist to design and build infrastructure for studying AI's societal impacts. You'll work with machine learning systems and data processing tools in San Francisco.

Ai Research Engineer
Anthropic is hiring a Research Engineer for their Frontier Red Team to focus on the safety of autonomous AI systems. You'll work on building and evaluating model organisms and developing defensive agents. This role requires expertise in AI capabilities research and security.