Senior GCP Data Engineer (Databricks) - Bułgaria, Rumunia

Remote job

Xebia

Leading global technology consultancy providing strategy, software engineering, advanced training, and managed services to help businesses thrive in the AI-enabled digital era.

View all jobs at Xebia

Apply now Apply later

Hello, let’s meet!

We are Xebia - a place where experts grow. For nearly two decades now, we've been developing digital solutions for clients from many industries and places across the globe. Among the brands we’ve worked with are UPS, McLaren, Aviva, Deloitte, and many, many more.

We're passionate about Cloud-based solutions. So much so, that we have a partnership with three of the largest Cloud providers in the business – Amazon Web Services (AWS), Microsoft Azure & Google Cloud Platform (GCP). We even became the first AWS Premier Consulting Partner in Poland.

Formerly we were known as PGS Software. In 2021, we joined Xebia Group – a family of interlinked companies driven by the desire to make a difference in the world of technology.

Xebia stands for innovation, talented team members, and technological excellence. Xebia means worldwide recognition, and thought leadership. This regularly provides us with the opportunity to work on global, innovative projects.

Our mission can be captured in one word: Authority. We want to be recognized as the authority in our field of expertise.

What makes us stand out? It's the little details, like our attitude, dedication to knowledge, and the belief in people's potential - emphasizing every team members development. Obviously, these things are not easy to present on paper – so make sure to visit us to see it with your own eyes!

Now, we've talked a lot about ourselves – but we'd love to hear more about you.

Send us your resume to start the conversation and join the #Xebia.

You will be:

  • responsible for at-scale infrastructure design, build and deployment with a focus on distributed systems,
  • building and maintaining architecture patterns for data processing, workflow definitions, and system to system integrations using Big Data and Cloud technologies,
  • evaluating and translating technical design to workable technical solutions/code and technical specifications at par with industry standards,
  • driving creation of re-usable artifacts,
  • establishing scalable, efficient, automated processes for data analysis, data model development, validation, and implementation,
  • working closely with analysts/data scientists to understand impact to the downstream data models,
  • writing efficient and well-organized software to ship products in an iterative, continual release environment,
  • contributing and promoting good software engineering practices across the team,
  • communicating clearly and effectively to technical and non-technical audiences,
  • defining data retention policies,
  • monitoring performance and advising any necessary infrastructure changes.

Requirements

Your profile:

  • 3+ years’ experience with GCP (BigQuery, Dataflow, Pub/Sub, Bigtable or other NoSQL database, Dataproc, Storage, Kubernetes Engine,
  • 5+ years’ experience with data engineering or backend/fullstack software development,
  • strong SQL skills,
  • Python scripting proficiency,
  • experience with data transformation tools - Databricks and Spark,
  • Data manipulation libraries (such as Pandas, NumPy, PySpark),
  • experience in structuring and modelling data in both relational and non-relational forms,
  • ability to elaborate and propose relational/non-relational approach,
  • normalization / denormalization and data warehousing concepts (star, snowflake schemas),
  • designing for transactional and analytical operations,
  • experience with CI/CD tooling (GitHub, Azure DevOps, Harness etc.),
  • good verbal and written communication skills in English.

Work from European Union region and work permit are required.

Nice to have:

  • Apache Hadoop,
  • experience with data modelling tools, preferably DBT,
  • Enterprise Data Warehouse solutions, preferably Snowflake,
  • familiarity with ETL tools (such as Informatica, Talend, Datastage, Stitch, Fivetran etc.),
  • experience in containerization and orchestration (Docker, Kubernetes etc.),
  • cloud (Azure, AWS, GCP) certification.


Recruitment Process:

CV review – HR call – Interview – optional Client Interview – Decision

Apply now Apply later

* Salary range is an estimate based on our AI, ML, Data Science Salary Index 💰

Job stats:  0  0  0
Category: Engineering Jobs

Tags: Architecture AWS Azure Big Data BigQuery Bigtable CI/CD Consulting Data analysis Databricks Dataflow Dataproc Data warehouse Data Warehousing dbt DevOps Distributed Systems Docker Engineering ETL FiveTran GCP GitHub Google Cloud Hadoop Informatica Kubernetes ML models NoSQL NumPy Pandas PySpark Python Snowflake Spark SQL Talend

Region: Remote/Anywhere

More jobs like this