OpenAI

About OpenAI

Empowering humanity through safe AI innovation

🏢 Tech👥 1001+ employees📅 Founded 2015📍 Mission District, San Francisco, CA💰 $68.9b4.2
B2CB2BArtificial IntelligenceEnterpriseSaaSAPIDevOps

Key Highlights

  • Headquartered in San Francisco, CA with 1,001+ employees
  • $68.9 billion raised in funding from top investors
  • Launched ChatGPT, gaining 1 million users in 5 days
  • 20-week paid parental leave and unlimited PTO policy

OpenAI is a leading AI research and development platform headquartered in the Mission District of San Francisco, CA. With over 1,001 employees, OpenAI has raised $68.9 billion in funding and is known for its groundbreaking products like ChatGPT, which gained over 1 million users within just five day...

🎁 Benefits

OpenAI offers flexible work hours and encourages unlimited paid time off, promoting at least 4 weeks of vacation per year. Employees enjoy comprehensi...

🌟 Culture

OpenAI's culture is centered around its mission to ensure that AGI benefits all of humanity. The company values transparency and ethical consideration...

Overview

OpenAI is hiring a Software Engineer for their Inference team to optimize and scale inference infrastructure on AMD GPU platforms. You'll work with technologies like Python, CUDA, and Triton. This position requires experience in distributed systems and performance optimization.

Job Description

Who you are

You have a strong background in software engineering with a focus on performance optimization and distributed systems — you've worked on scaling infrastructure and have experience with GPU platforms. Your expertise in Python and CUDA allows you to develop efficient solutions for complex problems. You thrive in collaborative environments, working closely with research and infrastructure teams to ensure high-performance execution of AI models.

You possess a deep understanding of model inference and have experience integrating internal model-serving infrastructure into GPU-backed systems. Your problem-solving skills enable you to debug and optimize distributed inference workloads across various layers, including memory, network, and compute. You are familiar with frameworks like Triton and have a passion for advancing inference performance on AMD accelerators.

What you'll do

In this role, you will own the bring-up, correctness, and performance of the OpenAI inference stack on AMD hardware. You will collaborate with partner teams to design and optimize high-performance GPU kernels, ensuring that our largest models run smoothly on new hardware. Your responsibilities will include validating the correctness, performance, and scalability of model execution on large GPU clusters, as well as integrating internal model-serving infrastructure into a variety of systems.

You will work across the stack, from low-level kernel performance to high-level distributed execution, and will be instrumental in shaping OpenAI's multi-platform inference capabilities. Your contributions will directly impact the efficiency and effectiveness of our AI models, empowering consumers, enterprises, and developers to leverage cutting-edge technology.

What we offer

At OpenAI, we believe in the potential of artificial intelligence to solve global challenges. We offer a collaborative work environment where you can grow your skills and make a significant impact. Join us in shaping the future of technology and enjoy the opportunity to work with some of the brightest minds in the field. We are committed to providing reasonable accommodations to applicants with disabilities and encourage you to apply even if your experience doesn't match every requirement.

Interested in this role?

Apply now or save it for later. Get alerts for similar jobs at OpenAI.

Similar Jobs You Might Like

Based on your interests and this role

OpenAI

Technical Lead

OpenAI📍 San Francisco

OpenAI is hiring a Technical Lead for the Sora team to optimize model serving efficiency and enhance inference performance. You'll work closely with research and product teams, leveraging your expertise in GPU and kernel-level systems.

Lead
10 months ago
OpenAI

Software Engineering

OpenAI📍 San Francisco - On-Site

OpenAI is hiring a Software Engineer for their Model Inference team to optimize AI models for high-volume production environments. You'll work with Azure and Python to enhance model performance and efficiency. This position requires 5+ years of experience in software engineering.

🏛️ On-SiteMid-Level
1 year ago
OpenAI

Software Engineering

OpenAI📍 San Francisco - On-Site

OpenAI is hiring a Software Engineer for their GPU Infrastructure team to ensure the reliability and uptime of their compute fleet. You'll work with cutting-edge technologies in a high-performance computing environment. This position requires experience in system-level investigations and automation.

🏛️ On-SiteMid-Level
2w ago
OpenAI

Software Engineering

OpenAI📍 San Francisco - On-Site

OpenAI is hiring a Senior Software Engineer for their Inference team to design and build a load balancer for AI model inference. You'll work with Python and distributed systems to ensure high reliability and performance. This role requires strong experience in debugging complex systems.

🏛️ On-SiteSenior
3 months ago
Together AI

Backend Engineer

Together AI📍 San Francisco - On-Site

Together AI is seeking a Senior Backend Engineer to build and optimize their Inference Platform for advanced generative AI models. You'll work with technologies like Python, Docker, and AWS to enhance performance and scalability. This role requires strong experience in backend engineering and machine learning.

🏛️ On-SiteSenior
1 month ago