Data Engineer / Devops
Lisbon, PT
Capgemini
A global leader in consulting, technology services and digital transformation, we offer an array of integrated services combining technology with deep sector expertise.About our company
At Capgemini Engineering, the world leader in engineering services, we bring together a global team of engineers, scientists, and architects to help the world’s most innovative companies unleash their potential. From autonomous cars to life-saving robots, our digital and software technology experts think outside the box as they provide unique R&D and engineering services across all industries. Join us for a career full of opportunities. Where you can make a difference. Where no two days are the same.
Your role
We are seeking a skilled Data Engineer to join our dynamic team. The ideal candidate will be responsible for developing, maintaining, and optimizing our data pipeline architecture while ensuring the availability and performance of critical data workflows. If you are passionate about data engineering, have hands-on experience with cloud platforms and big data technologies, and are a strong problem-solver, this is the role for you.
- Design, build, and maintain scalable, efficient, and reliable data pipelines to support data processing and analytics;
- Implement best practices in Big Data tools and frameworks, ensuring data integrity and performance;
- Collaborate with cross-functional teams to integrate new data sources and optimize data pipelines;
- Work with complex datasets, ensuring data quality, transformation, and storage in BigQuery;
- Troubleshoot data pipeline issues and implement long-term solutions to improve system stability;
- Write and optimize complex SQL queries for data extraction and analysis;
- Utilize Kafka for real-time data streaming and event-driven architecture;
- Implement solutions using Python for data processing and ETL workflows;
- Perform code reviews and maintain high-quality code standards.
Your profile
- Proficiency in BigQuery with 3+ years of experience;
- 3+ years of experience of Spark in large-scale data processing;
- 3+ years of experience of SQL in writing and optimizing complex queries;
- Advanced proficiency in Python with 4+ years of experience in scripting and automation;
- 2+ years of experience of Kafka with real-time data streaming;
- Effective communication in English, both verbal and written.
Nice to Have:
- Experience with Google Dataflow and Google Dataproc for managing data pipelines in GCP;
- Knowledge of Scala for parallel data processing;
- Familiarity with Hadoop for distributed storage and processing of big data.
What you'll love about working here
- Join a multicultural and inclusive team environment;
- Enjoy a supportive atmosphere promoting work-life balance;
- Engage in exciting national and international projects;
- Hybrid work;
- Your career growth is central to our mission. Our array of career growth programs and diverse professionals are crafted to support you in exploring a world of opportunities;
- Training and certifications programs;
- Referral program with bonuses for talent recommendations;
- Great office locations.
Why Capgemini?
Choosing Capgemini means choosing a company where you will be empowered to shape your career in the way you’d like, where you’ll be supported and inspired by a collaborative community of colleagues around the world, and where you’ll be able to reimagine what’s possible. Join us and help the world’s leading organizations unlock the value of technology and build a more sustainable, more inclusive world.
Apply now!
* Salary range is an estimate based on our AI, ML, Data Science Salary Index 💰
Tags: Architecture Big Data BigQuery Dataflow Data pipelines Dataproc Data quality DevOps Engineering ETL GCP Hadoop Kafka Pipelines Python R R&D Scala Spark SQL Streaming
Perks/benefits: Startup environment
More jobs like this
Explore more career opportunities
Find even more open roles below ordered by popularity of job title or skills/products/technologies used.