ML Research Engineer
Bay Area, CA
⚠️ We'll shut down after Aug 1st - try foo🦍 for all jobs in tech ⚠️
Why we exist: Oumi is on a mission to make frontier AI truly open for all. We are founded on the belief that AI will have a transformative impact on humanity, and that developing it collectively, in the open, is the best path forward to ensure that it is done efficiently and safely.
What we do: Oumi provides an all-in-one platform to build state-of-the-art AI models, end to end, from data preparation to production deployment, empowering innovators to build cutting-edge models at any scale. Oumi also develops open foundation models in collaboration with academic collaborators and the open community.
Our Approach: Oumi is fundamentally an open-source first company, with open-collaboration across the community as a core principle. Our work is:
Open Source First: All our platform and core technology is open source
Research-driven: We conduct and publish original research in AI, collaborating with our community of academic research labs and collaborators
Community-powered: We believe in the power of open-collaboration and welcome contributions from researchers and developers worldwide
We’re looking for a Research Engineer to join our team working on generative AI and LLMs. In this role, you'll bridge research and engineering—designing scalable infrastructure, enabling cutting-edge experiments, and helping open-source the next generation of LLMs. You will collbaborate closely with our research team and the open-source community to build tools, run evaluations, and contribute to models that are safe, performant, and accessible.
What you'll do:
Design and build systems to support training, fine-tuning, and evaluating large language models.
Partner with researchers to define experiments, write reusable code, run benchmarks, and interpret results.
Work on LLM alignment and tuning using techniques like reinforcement learning (RLHF), supervised fine-tuning, and prompt optimization.
Develop scalable ML pipelines for distributed training (e.g., across multi-GPU and multi-node environments).
Contribute to open-source tooling and models to support transparency and community collaboration.
Optimize performance across the ML stack—from data loading to deployment.
What you’ll bring:
Strong experience in machine learning, deep learning, or NLP—especially in generative AI or LLMs.
Solid programming skills in Python, and experience with ML frameworks like PyTorch.
Experience designing or maintaining ML infrastructure at scale (e.g., cloud-based training, distributed systems).
Comfort working in highly collaborative environments with research and engineering teams.
Bonus: experience with academic publications, open-source contributions, or LLM alignment work.
Share Oumi's values: Beneficial for all, Customer-obsessed, Radical Ownership, Exceptional Teammates, Science-grounded.
Competitive salary: $140,000 - $220,000
Equity in a high-growth startup
Comprehensive health, dental and vision insurance
21 days PTO
Regular team offsites and events
Tags: Deep Learning Distributed Systems Engineering Generative AI GPU LLMs Machine Learning ML infrastructure NLP Open Source Pipelines Python PyTorch Reinforcement Learning Research RLHF
Perks/benefits: Career development Competitive pay Equity / stock options Health care Salary bonus Startup environment Team events Transparency
More jobs like this
Explore more career opportunities
Find even more open roles below ordered by popularity of job title or skills/products/technologies used.