
About OpenAI
Empowering humanity through safe AI innovation
Key Highlights
- Headquartered in San Francisco, CA with 1,001+ employees
- $68.9 billion raised in funding from top investors
- Launched ChatGPT, gaining 1 million users in 5 days
- 20-week paid parental leave and unlimited PTO policy
OpenAI is a leading AI research and development platform headquartered in the Mission District of San Francisco, CA. With over 1,001 employees, OpenAI has raised $68.9 billion in funding and is known for its groundbreaking products like ChatGPT, which gained over 1 million users within just five day...
π Benefits
OpenAI offers flexible work hours and encourages unlimited paid time off, promoting at least 4 weeks of vacation per year. Employees enjoy comprehensi...
π Culture
OpenAI's culture is centered around its mission to ensure that AGI benefits all of humanity. The company values transparency and ethical consideration...

Technical Lead β’ Lead
OpenAI β’ San Francisco
Skills & Technologies
Overview
OpenAI is hiring a Technical Lead for the Sora team to optimize model serving efficiency and enhance inference performance. You'll work closely with research and product teams, leveraging your expertise in GPU and kernel-level systems.
Job Description
Who you are
You have deep expertise in model performance optimization, particularly at the inference layer β your experience includes driving initiatives that enhance system efficiency and throughput. You possess a strong background in kernel-level systems and data movement, which allows you to tackle complex performance tuning challenges effectively.
Your excitement about scaling high-performing AI systems that serve real-world, multimodal applications sets you apart β you understand the intricacies involved in integrating multimodal functionalities into AI products. You thrive in collaborative environments, partnering closely with research and product teams to ensure models perform effectively at scale.
What you'll do
In this role, you will perform engineering efforts focused on improving model serving and inference performance β your contributions will directly impact the reliability and scalability of the Sora teamβs initiatives. You will drive optimizations from a kernel and data movement perspective, enhancing system throughput and reliability.
You will design, build, and improve critical serving infrastructure to support Soraβs growth and reliability needs β your work will enable leadership to focus on higher-leverage initiatives by establishing a stronger technical foundation. Collaborating with researchers, you will assist in developing inference-friendly models that align with OpenAIβs mission of broad societal benefit.
What we offer
At OpenAI, you will be part of a pioneering team that is shaping the future of technology β we are committed to providing reasonable accommodations to applicants with disabilities. Join us in our mission to ensure the benefits of AI are widely shared and contribute to solving immense global challenges.
Interested in this role?
Apply now or save it for later. Get alerts for similar jobs at OpenAI.
Similar Jobs You Might Like
Based on your interests and this role

Software Engineering
OpenAI is hiring a Software Engineer for their Inference team to optimize and scale inference infrastructure on AMD GPU platforms. You'll work with technologies like Python, CUDA, and Triton. This position requires experience in distributed systems and performance optimization.

Software Engineering
OpenAI is hiring a Software Engineer for their Model Inference team to optimize AI models for high-volume production environments. You'll work with Azure and Python to enhance model performance and efficiency. This position requires 5+ years of experience in software engineering.

Ai Research Engineer
Together AI is hiring a Systems Research Engineer specialized in GPU Programming to develop and optimize GPU-accelerated kernels for ML/AI applications. You'll collaborate with cross-functional teams and leverage your expertise in GPU programming and parallel computing. This role requires a strong background in GPU programming techniques.

Software Engineering
OpenAI is hiring a Software Engineer for their GPU Infrastructure team to ensure the reliability and uptime of their compute fleet. You'll work with cutting-edge technologies in a high-performance computing environment. This position requires experience in system-level investigations and automation.

Gpu Kernel Engineer
Baseten is hiring a GPU Kernel Engineer to optimize performance for cutting-edge AI workloads. You'll work with C, C++, and CUDA in San Francisco. This position requires experience in low-level optimization and machine learning.