
About OpenAI
Empowering humanity through safe AI innovation
Key Highlights
- Headquartered in San Francisco, CA with 1,001+ employees
- $68.9 billion raised in funding from top investors
- Launched ChatGPT, gaining 1 million users in 5 days
- 20-week paid parental leave and unlimited PTO policy
OpenAI is a leading AI research and development platform headquartered in the Mission District of San Francisco, CA. With over 1,001 employees, OpenAI has raised $68.9 billion in funding and is known for its groundbreaking products like ChatGPT, which gained over 1 million users within just five day...
🎁 Benefits
OpenAI offers flexible work hours and encourages unlimited paid time off, promoting at least 4 weeks of vacation per year. Employees enjoy comprehensi...
🌟 Culture
OpenAI's culture is centered around its mission to ensure that AGI benefits all of humanity. The company values transparency and ethical consideration...
Skills & Technologies
Overview
OpenAI is hiring a Data Engineer to build and operate data infrastructure that supports massive compute fleets and storage systems. You'll work with technologies like Apache Spark, Kafka, and Airflow in San Francisco.
Job Description
Who you are
You have a strong background in data engineering with experience in building and operating large-scale data infrastructure — you've designed and implemented data lakes and metadata systems, ensuring they are reliable and efficient. Your expertise includes working with high-throughput streaming platforms, and you are familiar with tools like Kafka and Flink, which you have used to manage real-time data flows effectively.
You possess a solid understanding of orchestration tools such as Airflow, enabling you to automate complex data workflows. Your programming skills in Python allow you to develop robust data pipelines and integrate various data sources seamlessly. You are also knowledgeable about data governance and security practices, ensuring that data access is secure and compliant with regulations.
What you'll do
In this role, you will be responsible for designing, building, and operating the next generation of data infrastructure at OpenAI. You will scale and harden big data compute and storage platforms, ensuring they are optimized for high performance and reliability. Your work will involve enabling secure and governed data access for machine learning and analytics, which is crucial for supporting OpenAI's innovative projects.
You will take full lifecycle ownership of the data infrastructure, from architecture and implementation to production monitoring and incident response. Collaborating closely with cross-functional teams, you will help redefine how people interact with data, contributing to intelligent interfaces and AI-assisted workflows that enhance data accessibility and usability.
What we offer
At OpenAI, you will be part of a mission-driven team that is shaping the future of technology through artificial intelligence. We provide a collaborative and inclusive work environment where your contributions will have a significant impact. You will have opportunities for professional growth and development, working alongside some of the brightest minds in the industry. We encourage you to apply even if your experience doesn't match every requirement, as we value diverse perspectives and backgrounds.
Interested in this role?
Apply now or save it for later. Get alerts for similar jobs at OpenAI.
Similar Jobs You Might Like
Based on your interests and this role

Software Engineering
OpenAI is hiring a Software Engineer for their Data Infrastructure team to design and implement dataset infrastructure for next-generation training stacks. You'll work with technologies like Python, Docker, and AWS in San Francisco.

Software Engineering
Plaid is hiring a Senior Software Engineer for their Data Infrastructure team to scale data systems and maintain data integrity. You'll work with technologies like Apache Spark and Data Warehousing in San Francisco.

Data Engineer
Intercom is seeking a Data Infrastructure Engineer to build and maintain distributed systems that empower teams with data. You'll work with technologies like Python, SQL, and AWS to create end-to-end data workflows. This role requires experience in data engineering and a strong understanding of data platforms.

Data Engineer
Krea is hiring a Data Engineer to build distributed systems for processing large amounts of data. You'll work with technologies like Kubernetes, PyTorch, and Pandas. This position requires experience in data engineering and familiarity with machine learning pipelines.

Data Engineer
Apple is hiring a Data Engineer to develop services that support mission-critical information security capabilities. You'll work with Apache Spark and collaborate with engineering teams to solve complex data challenges. This position requires a strong background in large-scale software development.