ML Research Engineer

Bay Area, CA

⚠️ We'll shut down after Aug 1st - try foo🦍 for all jobs in tech ⚠️

Oumi

Building truly open, reliable frontier AI.

View all jobs at Oumi

Apply now Apply later

About Oumi

Why we exist: Oumi is on a mission to make frontier AI truly open for all. We are founded on the belief that AI will have a transformative impact on humanity, and that developing it collectively, in the open, is the best path forward to ensure that it is done efficiently and safely.

What we do: Oumi provides an all-in-one platform to build state-of-the-art AI models, end to end, from data preparation to production deployment, empowering innovators to build cutting-edge models at any scale. Oumi also develops open foundation models in collaboration with academic collaborators and the open community.

Our Approach: Oumi is fundamentally an open-source first company, with open-collaboration across the community as a core principle. Our work is:

  • Open Source First: All our platform and core technology is open source

  • Research-driven: We conduct and publish original research in AI, collaborating with our community of academic research labs and collaborators

  • Community-powered: We believe in the power of open-collaboration and welcome contributions from researchers and developers worldwide

Role Overview

We’re looking for a Research Engineer to join our team working on generative AI and LLMs. In this role, you'll bridge research and engineering—designing scalable infrastructure, enabling cutting-edge experiments, and helping open-source the next generation of LLMs. You will collbaborate closely with our research team and the open-source community to build tools, run evaluations, and contribute to models that are safe, performant, and accessible.

What you'll do:

  • Design and build systems to support training, fine-tuning, and evaluating large language models.

  • Partner with researchers to define experiments, write reusable code, run benchmarks, and interpret results.

  • Work on LLM alignment and tuning using techniques like reinforcement learning (RLHF), supervised fine-tuning, and prompt optimization.

  • Develop scalable ML pipelines for distributed training (e.g., across multi-GPU and multi-node environments).

  • Contribute to open-source tooling and models to support transparency and community collaboration.

  • Optimize performance across the ML stack—from data loading to deployment.

What you’ll bring:

  • Strong experience in machine learning, deep learning, or NLP—especially in generative AI or LLMs.

  • Solid programming skills in Python, and experience with ML frameworks like PyTorch.

  • Experience designing or maintaining ML infrastructure at scale (e.g., cloud-based training, distributed systems).

  • Comfort working in highly collaborative environments with research and engineering teams.

  • Bonus: experience with academic publications, open-source contributions, or LLM alignment work.

  • Share Oumi's values: Beneficial for all, Customer-obsessed, Radical Ownership, Exceptional Teammates, Science-grounded.

Benefits
  • Competitive salary: $140,000 - $220,000

  • Equity in a high-growth startup

  • Comprehensive health, dental and vision insurance

  • 21 days PTO

  • Regular team offsites and events

Apply now Apply later
Job stats:  2  0  0

Tags: Deep Learning Distributed Systems Engineering Generative AI GPU LLMs Machine Learning ML infrastructure NLP Open Source Pipelines Python PyTorch Reinforcement Learning Research RLHF

Perks/benefits: Career development Competitive pay Equity / stock options Health care Salary bonus Startup environment Team events Transparency

Regions: Remote/Anywhere North America
Country: United States

More jobs like this