Member of Technical Staff - Edge AI Inference Engineer
Boston
Liquid AI
We build capable and efficient general-purpose AI systems at every scale. Liquid Foundation Models (LFMs) are a new generation of generative AI models that achieve state-of-the-art performance at every scale, while maintaining a smaller memory...
Liquid AI, an MIT spin-off, is a foundation model company headquartered in Boston, Massachusetts. Our mission is to build capable and efficient general-purpose AI systems at every scale.
Our goal at Liquid is to build the most capable AI systems to solve problems at every scale, such that users can build, access, and control their AI solutions. This is to ensure that AI will get meaningfully, reliably and efficiently integrated at all enterprises. Long term, Liquid will create and deploy frontier-AI-powered solutions that are available to everyone.
What this role actually is:
As we prepare to deploy our models across various edge device types, including CPUs, embedded GPUs, and NPUs, we seek an expert to optimize inference stacks tailored to each platform. We're looking for someone who can take our models, dive deep into the task, and return with a highly optimized inference stack, leveraging existing frameworks like llama.cpp, Executorch, and TensorRT to deliver exceptional throughput and low latency.
The ideal candidate is a highly skilled engineer with extensive experience in inference on embedded hardware and a deep understanding of CPU, NPU, and GPU architectures. They should be self-motivated, capable of working independently, and driven by a passion for optimizing performance across diverse edge hardware platforms.
Proficiency in building and enhancing edge inference stacks is essential. Additionally, experience with mobile development and expertise in cache-aware algorithms will be highly valued.
Our goal at Liquid is to build the most capable AI systems to solve problems at every scale, such that users can build, access, and control their AI solutions. This is to ensure that AI will get meaningfully, reliably and efficiently integrated at all enterprises. Long term, Liquid will create and deploy frontier-AI-powered solutions that are available to everyone.
What this role actually is:
As we prepare to deploy our models across various edge device types, including CPUs, embedded GPUs, and NPUs, we seek an expert to optimize inference stacks tailored to each platform. We're looking for someone who can take our models, dive deep into the task, and return with a highly optimized inference stack, leveraging existing frameworks like llama.cpp, Executorch, and TensorRT to deliver exceptional throughput and low latency.
The ideal candidate is a highly skilled engineer with extensive experience in inference on embedded hardware and a deep understanding of CPU, NPU, and GPU architectures. They should be self-motivated, capable of working independently, and driven by a passion for optimizing performance across diverse edge hardware platforms.
Proficiency in building and enhancing edge inference stacks is essential. Additionally, experience with mobile development and expertise in cache-aware algorithms will be highly valued.
Responsibilities
- Strong ML Experience: Proficiency in Python and PyTorch to effectively interface with the ML team at a deeply technical level.
- Hardware Awareness: Must understand modern hardware architecture, including cache hierarchies and memory access patterns, and their impact on performance.
- Proficient in Coding: Expertise in Python, C++, or Rust for AI-driven real-time embedded systems
- Optimization of Low-Level Primitives: Responsible for optimizing core primitives to ensure efficient model execution.
- Self-Guided and Ownership: Ability to independently take a PyTorch model and inference requirements and deliver a fully optimized edge inference stack with minimal guidance
* Salary range is an estimate based on our AI, ML, Data Science Salary Index 💰
Job stats:
1
0
0
Categories:
Deep Learning Jobs
Engineering Jobs
Leadership Jobs
Tags: Architecture GPU LLaMA Machine Learning Python PyTorch Rust TensorRT
Region:
North America
Country:
United States
More jobs like this
Explore more career opportunities
Find even more open roles below ordered by popularity of job title or skills/products/technologies used.
BI Developer jobsSr. Data Engineer jobsData Engineer II jobsBusiness Intelligence Analyst jobsPrincipal Data Engineer jobsStaff Data Scientist jobsStaff Machine Learning Engineer jobsData Manager jobsData Science Manager jobsPrincipal Software Engineer jobsData Science Intern jobsBusiness Data Analyst jobsJunior Data Analyst jobsData Analyst Intern jobsData Specialist jobsSoftware Engineer II jobsLead Data Analyst jobsResearch Scientist jobsSr. Data Scientist jobsDevOps Engineer jobsStaff Software Engineer jobsAI/ML Engineer jobsData Engineer III jobsSenior Backend Engineer jobsBI Analyst jobs
Git jobsAirflow jobsEconomics jobsOpen Source jobsLinux jobsComputer Vision jobsKafka jobsGoogle Cloud jobsJavaScript jobsMLOps jobsNoSQL jobsData Warehousing jobsTerraform jobsPhysics jobsKPIs jobsRDBMS jobsPostgreSQL jobsScikit-learn jobsBanking jobsHadoop jobsScala jobsGitHub jobsData warehouse jobsStreaming jobsPandas jobs
Classification jobsR&D jobsBigQuery jobsDistributed Systems jobsOracle jobsPySpark jobsdbt jobsLooker jobsCX jobsScrum jobsReact jobsRAG jobsMicroservices jobsRobotics jobsJira jobsRedshift jobsIndustrial jobsSAS jobsData Mining jobsNumPy jobsPrompt engineering jobsGPT jobsELT jobsMySQL jobsData strategy jobs