
About Anthropic
Building safe and reliable AI systems for everyone
Key Highlights
- Headquartered in SoMa, San Francisco, CA
- Raised $29.3 billion in funding, including $13 billion Series F
- Over 1,000 employees focused on AI safety and research
- Launched Claude, an AI chat assistant rivaling ChatGPT
Anthropic, headquartered in SoMa, San Francisco, is an AI safety and research company focused on developing reliable, interpretable, and steerable AI systems. With over 1,000 employees and backed by Google, Anthropic has raised $29.3 billion in funding, including a monumental Series F round of $13 b...
🎁 Benefits
Anthropic offers comprehensive health, dental, and vision insurance for employees and their dependents, along with inclusive fertility benefits via Ca...
🌟 Culture
Anthropic's culture is rooted in AI safety and reliability, with a focus on producing less harmful outputs compared to existing AI systems. The compan...
Skills & Technologies
Overview
Anthropic is seeking a Research Scientist/Engineer for their Alignment Finetuning team to develop techniques for training language models aligned with human values. You'll work with Python to implement novel finetuning techniques and improve model behavior.
Job Description
Who you are
You have a strong background in AI and machine learning, with experience in developing and implementing novel techniques for model training. Your expertise in Python allows you to create advanced training pipelines and evaluation frameworks that measure alignment properties in models. You are committed to building AI systems that are safe and beneficial for society, demonstrating a deep understanding of moral reasoning and character in AI behavior.
You thrive in collaborative environments, working across teams to integrate alignment improvements into production models. Your ability to automate and scale processes is complemented by your innovative mindset, always seeking new ways to enhance model performance and alignment with human values. You are passionate about the ethical implications of AI and are eager to contribute to a mission-driven organization.
What you'll do
As a Research Scientist/Engineer at Anthropic, you will lead the development of techniques aimed at training language models that exhibit better alignment with human values. This includes creating and implementing novel finetuning techniques using synthetic data generation. You will also be responsible for developing evaluation frameworks to measure alignment properties in models, ensuring that they demonstrate improved honesty and character.
Collaboration is key in this role, as you will work closely with other researchers and engineers to integrate alignment improvements into production models. You will develop processes to help automate and scale the work of the team, contributing to the overall mission of creating reliable and interpretable AI systems. Your work will directly impact the safety and beneficial nature of AI technologies, making a meaningful difference in the field.
What we offer
Anthropic provides a supportive and innovative work environment, with competitive compensation and benefits. You will have the opportunity to work with a talented team of researchers and engineers dedicated to building beneficial AI systems. Flexible working hours and a collaborative office space in San Francisco enhance the work experience, allowing you to thrive in your role while contributing to a mission that prioritizes safety and ethical considerations in AI development.
Interested in this role?
Apply now or save it for later. Get alerts for similar jobs at Anthropic.
Similar Jobs You Might Like
Based on your interests and this role

Ai Research Engineer
OpenAI is hiring a Researcher in Alignment to ensure AI systems follow human intent in complex scenarios. You'll focus on designing scalable solutions for AI alignment. This role is based in San Francisco.

Ai Research Engineer
Anthropic is hiring a Research Engineer / Scientist in Alignment Science to build and run machine learning experiments focused on AI safety. You'll work with technologies like Python and contribute to exploratory research on powerful AI systems. This role requires a blend of scientific and engineering skills.

Ai Research Engineer
Anthropic is hiring a Research Engineer/Scientist in Alignment Science to conduct machine learning experiments focused on AI safety. You'll work on AI control and alignment stress-testing in London.

Research Scientist
Anthropic is seeking a Research Scientist/Engineer focused on honesty to develop techniques that minimize hallucinations and enhance truthfulness in language models. You'll work primarily with Python to create robust AI systems. This role is based in New York City or San Francisco.

Ai Research Engineer
Anthropic is hiring a Staff Research Engineer for their Discovery Team to work on developing advanced AI systems. You'll focus on identifying and addressing key blockers in the path to scientific AGI, utilizing skills in Python, Docker, and machine learning. This role requires familiarity with distributed systems and performance optimization.