
About OpenAI
Empowering humanity through safe AI innovation
Key Highlights
- Headquartered in San Francisco, CA with 1,001+ employees
- $68.9 billion raised in funding from top investors
- Launched ChatGPT, gaining 1 million users in 5 days
- 20-week paid parental leave and unlimited PTO policy
OpenAI is a leading AI research and development platform headquartered in the Mission District of San Francisco, CA. With over 1,001 employees, OpenAI has raised $68.9 billion in funding and is known for its groundbreaking products like ChatGPT, which gained over 1 million users within just five day...
🎁 Benefits
OpenAI offers flexible work hours and encourages unlimited paid time off, promoting at least 4 weeks of vacation per year. Employees enjoy comprehensi...
🌟 Culture
OpenAI's culture is centered around its mission to ensure that AGI benefits all of humanity. The company values transparency and ethical consideration...
Skills & Technologies
Overview
OpenAI is hiring a Software Engineer for their Inference team to build and optimize infrastructure for multimodal models. You'll work with technologies like Python, TensorFlow, and Kubernetes to serve real-time audio and image workloads. This position requires experience in machine learning and software engineering.
Job Description
Who you are
You have a strong background in software engineering with a focus on building scalable systems — your experience includes working with machine learning models and optimizing infrastructure for high-performance applications. You are proficient in Python and have hands-on experience with frameworks like TensorFlow and PyTorch, which you have used to develop and deploy machine learning models in production environments.
You understand the complexities of multimodal systems and have experience collaborating with cross-functional teams, including researchers and product managers — your ability to communicate technical concepts clearly helps bridge the gap between engineering and research. You thrive in fast-paced environments and are eager to push the boundaries of what AI can achieve.
What you'll do
As a Software Engineer on the Inference team, you will design and implement infrastructure for serving large-scale multimodal models — your work will ensure that OpenAI's advanced models are performant and scalable in production. You will collaborate closely with researchers to integrate new modalities of interaction, optimizing systems for high-throughput and low-latency performance.
You will be responsible for building reliable systems that handle real-time audio, image, and other multimodal workloads — this includes optimizing data pipelines and ensuring that the infrastructure can support diverse model sizes and interactions. Your role will involve continuous improvement of the deployment processes, ensuring compliance with best practices in software engineering and machine learning.
You will also participate in code reviews and contribute to the overall technical direction of the team — mentoring junior engineers and sharing your knowledge will be key aspects of your role. You will have the opportunity to work on cutting-edge AI technologies and contribute to projects that have a significant impact on the future of technology.
What we offer
At OpenAI, you will be part of a mission-driven team that is dedicated to advancing artificial intelligence for the benefit of humanity — we offer a collaborative and inclusive work environment where innovation is encouraged. You will have access to resources and support to help you grow in your career, including opportunities for professional development and training.
We believe in the importance of work-life balance and offer flexible working arrangements to accommodate your needs — our office in San Francisco is equipped with modern amenities to support a productive work environment. Join us in shaping the future of AI and technology.
Interested in this role?
Apply now or save it for later. Get alerts for similar jobs at OpenAI.
Similar Jobs You Might Like
Based on your interests and this role

Software Engineering
OpenAI is hiring a Software Engineer for their Model Inference team to optimize AI models for high-volume production environments. You'll work with Azure and Python to enhance model performance and efficiency. This position requires 5+ years of experience in software engineering.

Staff Engineer
Cohere is hiring a Staff Software Engineer for their Inference Infrastructure team to build high-performance AI platforms. You'll work with technologies like Python and Docker to deploy optimized NLP models. This role requires experience in machine learning and scalable systems.

Software Engineering
OpenAI is hiring a Senior Software Engineer for their Inference team to design and build a load balancer for AI model inference. You'll work with Python and distributed systems to ensure high reliability and performance. This role requires strong experience in debugging complex systems.

Technical Lead
OpenAI is hiring a Technical Lead for the Sora team to optimize model serving efficiency and enhance inference performance. You'll work closely with research and product teams, leveraging your expertise in GPU and kernel-level systems.

Backend Engineer
Together AI is seeking a Senior Backend Engineer to build and optimize their Inference Platform for advanced generative AI models. You'll work with technologies like Python, Docker, and AWS to enhance performance and scalability. This role requires strong experience in backend engineering and machine learning.