GenAI Engineer

Warszawa, Województwo mazowieckie, PL

Tango

Reward Yourself! Join the Tango Team. We’re doing incredible things with awesome people, and we’d love for you to join us!

View all jobs at Tango

Apply now Apply later

Description

We are looking for an GenAI Engineer to drive the development of AI-powered conversation agents, real-time chat moderation systems and generative content solutions within the Tango application and related projects.


Responsibilities

  • Design, develop, and optimize large language models (LLMs) and generative AI solutions for various applications, ensuring performance scalability and real-time response capabilities.
  • Fine-tune and customize open-source and proprietary LLMs for domain-specific applications.
  • Implement RLHF and other model improvement techniques.
  • Optimize model inference, latency, and memory footprint for scalability and efficiency.
  • Build LLM-powered AI agents capable of autonomous decision-making and multi-step reasoning.
  • Implement memory-augmented and multi-agent architectures for long-term task execution.
  • Design, test, and optimize prompts for various AI-driven applications.
  • Develop and evaluate prompt strategies for improved model performance and cost-efficiency.
  • Work with APIs, on-premises, and cloud-based LLM hosting solutions.


Collaboration & Research

  • Conduct research to stay on top of emerging trends in AI at large.
  • Research and experiment with SOTA (State-of-the-Art) AI techniques, architectures, and frameworks
  • Collaborate with cross-functional teams to understand project requirements, align AI solutions with business objectives and integrate models into real-world applications.
  • Contribute to open-source projects, research papers, or internal documentation.

Requirements

  • Bachelor’s/Master’s/Ph.D. in Computer Science, Machine Learning, Artificial Intelligence, or a related field.
  • Proven experience as an AI Engineer, Machine Learning Engineer, or similar role with a focus on generative models and retrieval-augmented techniques.
  • Experience in fine-tuning and model compression techniques.
  • Proficiency in Python and AI frameworks like PyTorch, TensorFlow.
  • Familiarity with open-source LLM models (Llama, Falcon, Mistral, GPT-NeoX, etc.).
  • Hands-on experience with Hugging Face, LangChain, LlamaIndex, OpenAI APIs, or similar ecosystems.
  • Knowledge of vector databases (FAISS, Pinecone, Weaviate, etc.) and retrieval-augmented generation (RAGpipelines).
  • Experience with autonomous AI agents and self-improving LLM architectures.
  • Experience with multi-modal models (e.g., integrating text, images, or other data forms).
  • Knowledge of LoRA, QLoRA, and PEFT (Parameter Efficient Fine-Tuning) and other cost-effective model adaptation techniques.
  • Understanding of hallucination reduction and fact-checking techniques for LLMs.
  • Excellent problem-solving skills, analytical thinking, and attention to detail.
  • Ability to communicate complex technical concepts to both technical and non-technical stakeholders.


What we offer:

  • Stock options grant (we’re a Silicon Valley Company)
  • Competitive salary
  • On-site position with 4 days from office / 1 day from home (no remote)
  • Medical insurance for you and 75% off for your relatives
  • Free lunches and dinners
  • Parking
  • Multisport card
  • Cheerful team spirit and fun office atmosphere

If this sounds like you, apply and help empower live entertainers and creators to build independent businesses around their live talents.

#LI-Onsite

Apply now Apply later

* Salary range is an estimate based on our AI, ML, Data Science Salary Index 💰

Job stats:  1  0  0

Tags: APIs Architecture Computer Science FAISS Generative AI Generative modeling GPT LangChain LLaMA LLMs LoRA Machine Learning Model inference OpenAI Open Source Pinecone Pipelines Python PyTorch RAG Research RLHF TensorFlow Weaviate

Perks/benefits: Competitive pay Equity / stock options

Region: Europe
Country: Poland

More jobs like this