Senior Machine Learning Engineer - LLM Systems & Research

Sofia, Bulgaria

⚠️ We'll shut down after Aug 1st - try foo🦍 for all jobs in tech ⚠️

Constructor TECH

Constructor is an all-in-one platform for education and research. Constructor elevates learning experiences, empowers educators and drives research breakthroughs.

View all jobs at Constructor TECH

Apply now Apply later

Our mission

Constructor’s mission is to enable all educational organisations to provide high-quality digital education to 10x people with 10x efficiency. 

With strong expertise in machine intelligence and data science, Constructor’s all-in-one platform for education and research addresses today’s pressing educational challenges: access inequality, tech clutter, and low engagement of students.

Our headquarters is located in 🇨🇭Switzerland, and we also have legal entities in 🇩🇪Germany, 🇧🇬Bulgaria, 🇷🇸Serbia, 🇹🇷Turkey, and 🇸🇬Singapore

Please send your resume in English only.

We are seeking an exceptional Senior Machine Learning Engineer with deep expertise in Large Language Model systems and applied research in educational AI. The successful candidate will be responsible for designing and implementing enterprise-scale LLM infrastructure, contributing to research initiatives, and building systems that advance educational technology.

Project Overview

You will be designing and implementing next-generation LLM systems that serve thousands of users across research institutions and educational organizations globally. This includes building distributed ML platforms, implementing research innovations, and creating LLM applications that enhance educational experiences.

Responsibilities

  1. LLM Architecture & Engineering: Design and implement large-scale LLM systems, establish technical standards for model deployment, and drive strategic technical decisions for ML infrastructure.
  2. Platform Engineering: Build foundational ML platforms that enable rapid experimentation, deployment, and scaling of LLM applications. Design systems for multi-tenancy, global distribution, and enterprise-grade reliability.
  3. Advanced MLOps & Infrastructure: Develop sophisticated ML infrastructure including automated training pipelines, multi-model serving architectures, advanced monitoring systems, and cost optimization strategies for large-scale deployments.
  4. Research & Innovation: Execute research roadmaps in educational AI, collaborate with academic institutions, and translate cutting-edge research into production systems. Contribute to publications and patent applications.
  5. Performance & Scale Engineering: Build systems capable of serving models at massive scale (10M+ requests/day), implement advanced optimization techniques, and design fault-tolerant distributed systems with sub-100ms latency requirements.
  6. Educational AI Development: Create LLM-powered applications specifically designed for educational use cases, implement learning analytics, and build systems that enhance teaching and learning experiences.

Required Qualifications

Educational Background

Master’s degree in Computer Science, Machine Learning, or related field, or Bachelor’s degree with equivalent experience. PhD preferred for research contributions.

Professional Experience

  • 5+ years of experience in Machine Learning Engineering with focus on production systems
  • 2+ years of hands-on experience building and scaling LLM systems in production environments
  • Experience building distributed systems serving millions of users

Advanced Engineering Skills

  • Systems Programming: Expert-level proficiency in Python, with experience in Rust or C++
  • ML Infrastructure: Extensive experience with ML platforms, model serving architectures, and distributed training systems (Ray, Horovod, DeepSpeed)
  • Cloud Architecture: Deep expertise in cloud-native architectures and multi-region deployments
  • Performance Engineering: Proven track record optimizing large-scale systems for latency and cost efficiency

LLM & Research Expertise

  • Advanced LLM Technologies: Expert-level experience with model optimization, quantization, and efficient inference techniques
  • Distributed Inference: Hands-on experience with model parallelism, pipeline parallelism, and distributed serving architectures
  • Research Integration: Experience implementing research papers into production systems and contributing to open-source ML projects

MLOps & Platform Skills

  • Platform Design: Experience building ML experimentation frameworks, model registries, and deployment automation
  • Advanced Monitoring: Expertise in ML observability systems and automated performance monitoring
  • Cost Optimization: Experience optimizing infrastructure costs for large-scale ML workloads

Preferred Qualifications

  1. Research Background: PhD in ML/AI with publications in top-tier venues (NeurIPS, ICML, ICLR, ACL), or equivalent industry research experience
  2. Enterprise Scale: Experience serving models at massive scale (>10M requests/day) with enterprise SLA requirements
  3. Specialized Hardware: Deep expertise in GPU optimization, custom hardware acceleration (TPUs, FPGAs), and emerging ML accelerators
  4. Open Source Contributions: Significant contributions to major ML frameworks or maintainer status in relevant repositories
  5. Educational AI Expertise: Domain knowledge in educational technology, learning sciences, or academic research environments

What You’ll Build

  • Advanced LLM Systems: Design and implement cutting-edge LLM serving infrastructure with enterprise-grade reliability and performance
  • Research Infrastructure: Develop platforms enabling rapid experimentation and seamless research-to-production workflows
  • Educational AI Applications: Build innovative LLM-powered tools that transform teaching and learning experiences
  • Performance Innovation: Create novel optimization techniques and serving architectures that push the boundaries of LLM efficiency

Technical Environment

Core Technology Stack

  • Languages: Python, Rust, C++, CUDA, SQL
  • ML Stack: PyTorch, JAX, Triton, Custom CUDA kernels, Distributed training frameworks
  • Infrastructure: Kubernetes, Terraform, Multi-cloud architectures
  • Data Systems: Distributed databases, Real-time streaming (Kafka), Vector databases
  • Monitoring: ML observability platforms, Custom metrics systems

What We Offer

  • 💻 Choice of work equipment (e.g., laptop, monitor, etc.)
  • 🇬🇧 English classes (iTalki – $130 monthly)
  • ⏰ Flexible schedule (we usually work between 09:00/10:00 and 18:00/19:00 CET or EET)
  • 👶 Newborn bonus (€500 per child)
  • 🧠 Patent remuneration
  • 🌴 Paid leave
  • 🧑‍💻 Remote work in locations without our offices
  • Hybrid work in locations with offices (2 days in-office, 3 days remote):
    • 🇧🇬 Sofia: 59 G. M. Dimitrov Blvd., NV Tower, 8th floor, 1700
    • 🇷🇸 Belgrade: Makedonska 12, 11000 Belgrade, Serbia
    • 🇹🇷 Istanbul: Rüzgarlı Bahçe Mah., Kavak Sok., Smart Plaza B Blok 31/B, 34805 Kavacık-Beykoz/İstanbul
    • 🇹🇷 Sakarya: Esentepe Mh., Akademiyolu Sk., Teknoloji Geliştirme Bölgesi No. 10 D/206, Serdivan, Sakarya
    • 🇹🇷 Izmir: Ege Üniversitesi Kampüsü, Erzene Mah., Ankara Cad., No:172/67, 35100 Bornova/İzmir

Constructor fosters equal opportunity for people of all backgrounds and identities. We are led by a gender-balanced board committed to building a diverse and inclusive organisation where everyone can become their best self. We do not discriminate based on age, disability, gender identity, sexual orientation, ethnicity, race, religion or belief, parental and family status, or other protected characteristics. We welcome applications from women, men and non-binary candidates of all ethnicities and socio-economic backgrounds. We encourage people belonging to underrepresented groups to apply.

Apply now Apply later

* Salary range is an estimate based on our AI, ML, Data Science Salary Index 💰

Job stats:  1  0  0

Tags: Architecture CAD Computer Science CUDA Distributed Systems Engineering GPU Horovod ICLR ICML JAX Kafka Kubernetes LLMs Machine intelligence Machine Learning ML infrastructure MLOps Model deployment NeurIPS Open Source PhD Pipelines Python PyTorch Research Rust SQL Streaming Teaching Terraform

Perks/benefits: Flex hours Gear Parental leave

Region: Europe
Country: Bulgaria

More jobs like this