
About Liquid AI
Efficient AI for a smarter future
Key Highlights
- Headquartered in Cambridge, Massachusetts
- 200+ employees with expertise in AI/ML
- Specializes in Liquid Foundation Models (LFMs)
- Focus on reducing computational overhead for AI solutions
Liquid AI, headquartered in Cambridge, Massachusetts, specializes in general-purpose artificial intelligence through its innovative Liquid Foundation Models (LFMs). These models are designed to deliver high performance while significantly reducing memory and computing resource requirements, making a...
🎁 Benefits
Employees enjoy competitive salaries, equity options, generous PTO policies, and opportunities for remote work. Liquid AI also offers a learning budge...
🌟 Culture
Liquid AI fosters a culture of innovation and efficiency, focusing on optimizing AI capabilities while minimizing resource usage. The company values c...
Skills & Technologies
Overview
Liquid AI is hiring a Member of Technical Staff - Edge Inference Engineer to optimize machine code for AI systems on resource-constrained devices. You'll work with technologies like Python and TensorFlow, requiring a deep understanding of ML architectures and hardware constraints.
Job Description
Who you are
You have a strong background in machine learning and systems engineering, with experience in optimizing code for resource-constrained devices. You understand the intricacies of ML architectures and can diagnose performance bottlenecks effectively. Your ability to work autonomously means you can tackle complex problems without needing constant guidance, and you thrive in high-ownership roles where your contributions directly impact production systems.
You possess a deep understanding of hardware-level optimizations, including cache hierarchies and memory access patterns. This knowledge allows you to reason about code performance and implement effective solutions. You are comfortable bridging the gap between machine learning and systems engineering, ensuring that your code runs efficiently on various devices, from smartphones to embedded systems.
What you'll do
As a Member of Technical Staff on the Edge Inference team, you will compile Liquid Foundation Models into optimized machine code that runs on devices with limited resources. You will collaborate closely with the technical lead to address challenges that require a nuanced understanding of both machine learning and hardware constraints. Your work will involve diagnosing performance issues, prototyping solutions, and iterating on your designs until you meet performance goals.
You will be responsible for shipping code to production that enhances model performance on real devices, ensuring low latency and minimal memory usage. Your contributions will be critical in making efficient on-device AI a reality, impacting various sectors including consumer electronics and automotive industries. You will also have the opportunity to work with cutting-edge technologies and contribute to open-source projects like llama.cpp, further enhancing your skills and professional growth.
What we offer
Liquid AI provides a competitive base salary along with equity in a rapidly growing company. We cover 100% of medical, dental, and vision premiums for employees and their dependents. Our financial benefits include a 401(k) matching program up to 4% of base pay. We also offer unlimited PTO and company-wide Refill Days to ensure a healthy work-life balance. Join us in shaping the future of AI technology while enjoying a supportive and innovative work environment.
Interested in this role?
Apply now or save it for later. Get alerts for similar jobs at Liquid AI.
Similar Jobs You Might Like
Based on your interests and this role

Staff Engineer
Cohere is hiring a Staff Software Engineer for their Inference Infrastructure team to build high-performance AI platforms. You'll work with technologies like Python and Docker to deploy optimized NLP models. This role requires experience in machine learning and scalable systems.

Backend Engineer
Together AI is seeking a Senior Backend Engineer to build and optimize their Inference Platform for advanced generative AI models. You'll work with technologies like Python, Docker, and AWS to enhance performance and scalability. This role requires strong experience in backend engineering and machine learning.

Technical Lead
OpenAI is hiring a Technical Lead for the Sora team to optimize model serving efficiency and enhance inference performance. You'll work closely with research and product teams, leveraging your expertise in GPU and kernel-level systems.

Ai Engineer
Cartesia is hiring an Inference Engineer to design and build low latency, scalable, and reliable model inference and serving stacks for cutting-edge foundation models. You'll work with Transformers and SSMs in San Francisco.

Ai Research Engineer
Liquid AI is hiring a Member of Technical Staff - ML Research Engineer to build vision-language models that run efficiently on-device. You'll work with Python, TensorFlow, and PyTorch to design and implement large-scale training experiments. This position requires expertise in machine learning and computer vision.