Principal AI Engineer

Bangalore - Carina, India

Red Hat

Red Hat is the world’s leading provider of enterprise open source solutions, including high-performing Linux, cloud, container, and Kubernetes technologies.

View all jobs at Red Hat

Apply now Apply later

About the Job:

The Data Development Insights & Strategy (DDIS) team is seeking a Principal AI Engineer to lead the design, development, and optimization of AI model lifecycle frameworks within Red Hat’s OpenShift AI and RHEL AI infrastructures. As a Principal AI Engineer, you will play a key leadership role in overseeing the strategic direction of AI model deployment and lifecycle management, collaborating across teams to ensure seamless integration, scalability, and performance of mission-critical AI models.

In this role, you will drive the development of innovative solutions for the AI model lifecycle, applying your deep expertise in MLOps/LLMOps, cloud computing, and distributed systems. You will be a technical leader who mentors and guides teams in collaboration with Products & Global Engineering (P&GE) and IT AI Infra to ensure efficient model deployment and maintenance in secure, scalable environments. This is an exciting opportunity for someone who wants to take a leadership role in influencing the strategic direction of Red Hat's AI innovations, driving the innovation and optimization of AI models and technologies.

What you will do?

  • Lead the design and development of scalable, efficient, and secure AI model lifecycle frameworks within Red Hat’s OpenShift and RHEL AI infrastructures, ensuring models are deployed and maintained with minimal disruption and optimal performance.

  • Define and implement the strategy for optimizing AI model deployment, scaling, and integration across hybrid cloud environments (AWS, GCP, Azure), working with cross-functional teams to ensure consistent high availability and operational excellence.

  • Spearhead the creation and optimization of CI/CD pipelines and automation for AI model deployments, leveraging tools such as Git, Jenkins, and Terraform, ensuring zero disruption during updates and integration.

  • Champion the use of advanced monitoring tools (e.g., OpenLLMetry, Splunk, Catchpoint) to monitor and optimize model performance, responding to issues and leading the troubleshooting of complex problems related to AI and LLM models.

  • Lead cross-functional collaboration in collaboration with Products & Global Engineering (P&GE) and IT AI Infra teams to ensure seamless integration of new models or model updates into production systems, adhering to best practices and minimizing downtime.

  • Define and oversee the structured process for handling feature requests (RFEs), prioritization, and resolution, ensuring transparency and timely delivery of updates and enhancements.

  • Lead and influence the adoption of new AI technologies, tools, and frameworks to ensure that Red Hat remains at the forefront of AI and machine learning advancements.

  • Drive performance improvements, model updates, and releases on a quarterly basis, ensuring RFEs are processed and resolved within agreed-upon timeframes and driving business adoption.

  • Oversee the fine-tuning and enhancement of large-scale models, including foundational models like Mistral and LLama, ensuring the optimal allocation of computational resources (GPU management, cost management strategies).

  • Lead a team of engineers, mentoring junior and senior talent, fostering an environment of collaboration and continuous learning, and driving the technical growth of the team.

  • Contribute to strategic discussions with leadership, influencing the direction of AI initiatives and ensuring alignment with broader business goals and technological advancements.

What you will bring?

  • A bachelor’s or master’s degree in Computer Science, Data Science, Machine Learning, or a related technical field is required.

  • Hands-on experience and demonstrated leadership in AI engineering and MLOps will be considered in lieu of formal degree requirements.

  • 10+ years of experience in AI or MLOps, with at least 3 years in a technical leadership role managing the deployment, optimization, and lifecycle of large-scale AI models. You should have deep expertise in cloud platforms (AWS, GCP, Azure) and containerized environments (OpenShift, Kubernetes), with a proven track record in scaling and managing AI infrastructure in production.

  • Experience optimizing large-scale distributed AI systems, automating deployment pipelines using CI/CD tools like Git, Jenkins, and Terraform, and leading performance monitoring using tools such as OpenLLMetry, Splunk, or Catchpoint. You should have a strong background in GPU-based computing and resource optimization (e.g., CUDA, MIG, vLLM) and be comfortable with high-performance computing environments.

  • Your leadership skills will be key, as you will mentor and guide engineers while fostering a collaborative, high-performance culture. You should also have a demonstrated ability to drive innovation, solve complex technical challenges, and work cross-functionally with teams to deliver AI model updates that align with evolving business needs. A solid understanding of Agile development processes and excellent communication skills are essential for this role.

  • Lastly, a passion for AI, continuous learning, and staying ahead of industry trends will be vital to your success at Red Hat.

Desired skills:

  • 10+ years of experience in AI, MLOps, or related fields, with a substantial portion of that time spent in technical leadership roles driving the strategic direction of AI infrastructure and model lifecycle management.

  • Extensive experience with foundational models such as Mistral, LLama, GPT, and their deployment, tuning, and scaling in production environments.

  • Proven ability to influence and drive AI and MLOps roadmaps, shaping technical strategy and execution in collaboration with senior leadership.

  • In-depth experience with performance monitoring, resource optimization, and troubleshooting of AI models in complex distributed environments.

  • Strong background in high-performance distributed systems and container orchestration, particularly in AI/ML workloads.

  • Proven experience in guiding and mentoring engineering teams to build high-performance capabilities, fostering a culture of continuous improvement and technical innovation.

As a Principal AI Engineer at Red Hat, you will have the opportunity to drive major strategic AI initiatives, influence the future of AI infrastructure, and lead a high-performing engineering team. This is a unique opportunity for a seasoned AI professional to shape the future of AI model lifecycle management at scale. If you’re ready to take on a technical leadership role with a high level of responsibility and impact, we encourage you to apply.

About Red Hat

Red Hat is the world’s leading provider of enterprise open source software solutions, using a community-powered approach to deliver high-performing Linux, cloud, container, and Kubernetes technologies. Spread across 40+ countries, our associates work flexibly across work environments, from in-office, to office-flex, to fully remote, depending on the requirements of their role. Red Hatters are encouraged to bring their best ideas, no matter their title or tenure. We're a leader in open source because of our open and inclusive environment. We hire creative, passionate people ready to contribute their ideas, help solve complex problems, and make an impact.

Diversity, Equity & Inclusion at Red Hat
Red Hat’s culture is built on the open source principles of transparency, collaboration, and inclusion, where the best ideas can come from anywhere and anyone. When this is realized, it empowers people from diverse backgrounds, perspectives, and experiences to come together to share ideas, challenge the status quo, and drive innovation. Our aspiration is that everyone experiences this culture with equal opportunity and access, and that all voices are not only heard but also celebrated. We hope you will join our celebration, and we welcome and encourage applicants from all the beautiful dimensions of diversity that compose our global village.

Equal Opportunity Policy (EEO)
Red Hat is proud to be an equal opportunity workplace and an affirmative action employer. We review applications for employment without regard to their race, color, religion, sex, sexual orientation, gender identity, national origin, ancestry, citizenship, age, veteran status, genetic information, physical or mental disability, medical condition, marital status, or any other basis prohibited by law.


Red Hat does not seek or accept unsolicited resumes or CVs from recruitment agencies. We are not responsible for, and will not pay, any fees, commissions, or any other payment related to unsolicited resumes or CVs except as required in a written contract between Red Hat and the recruitment agency or party requesting payment of a fee.


Red Hat supports individuals with disabilities and provides reasonable accommodations to job applicants. If you need assistance completing our online job application, email application-assistance@redhat.com. General inquiries, such as those regarding the status of a job application, will not receive a reply.

Apply now Apply later

* Salary range is an estimate based on our AI, ML, Data Science Salary Index 💰

Job stats:  0  0  0

Tags: Agile AWS Azure CI/CD Computer Science CUDA Distributed Systems Engineering GCP Git GPT GPU Jenkins Kubernetes Linux LLaMA LLMOps LLMs Machine Learning ML infrastructure MLOps Model deployment Open Source Pipelines Splunk Terraform vLLM

Perks/benefits: Career development Transparency

Regions: Remote/Anywhere Asia/Pacific
Country: India

More jobs like this