DataOps Remote

Porto, Portugal

Alter Solutions

Looking for an IT partner? We're an IT consulting company with expertise in Nearshore software development and Cybersecurity. Offices in 13 countries

View all jobs at Alter Solutions

Apply now Apply later

Job Description

MAIN RESPONSABILITIES
• Incident Response
o Understand problems from a user perspective and communicate to clearly 
understand the issue.
o Reproduce bugs or issues that users are facing.
o Apply root cause analysis to quickly and efficiently 
o Find the root cause of the problem, patch it, test it, and communicate with the 
end user.
o Write postmortems summarizing every step of resolution and helping the team 
to track all issues.
o Monitor existing flows and infrastructure and perform the same tasks when 
discovering bugs/issues through monitoring and alerting.
• Maintenance
o Monitor flows and infrastructure to identify potential issues.
o Adapt configurations to keep flows and infrastructures working as expected, 
keeping the operations without incident.
• Database Optimization
o Track costs and time of processing through dedicated dashboards.
o Alert people who query tables the wrong way, involving high costs.
o Track down jobs, views, and tables that are running inefficiently and occur 
either high costs or low speed of execution.
o Optimize jobs, queries, and tables to optimize both costs and speed of 
execution.
• Infrastructure Management
o Manage infrastructure through Terraform.
o Share and propose good practices.
o Decommission useless infrastructures such as services, tables, or virtual 
machines.
• Deployments
o Track future deployments with a Data Architect and participate in Deployment 
Reviews.
o Share and propose good practices of deployment.
o Accompany Data Engineers during the entire process of deployments.
o Accompany Data Engineers in the following period of active monitoring.
o Ensure diligent application of deployment process, logging, and monitoring 
strategy.
o Take over newly deployed flows in the run process.
REQUESTED HARD SKILLS
• Google Cloud Platform: General knowledge of the platform and various services, 
and at least one year of experience with GCP.
• Apache Airflow: At least two years of experience with the Airflow orchestrator, 
experience with Google Composer is a plus.
• Google BigQuery: Extensive experience (at least 4 years) with GBQ, know how to 
optimize tables and queries, and able to design database architecture.
• Terraform: At least two years of experience with Terraform, and know good 
practices of GitOps.
• Apache Spark: this is an optional expertise we would value. Some of our pipelines 
use pySpark. 
• Additional Knowledge and Experience that are a Plus:
o Pub/Sub
o Kafka
o Azure Analysis Services
o Google Cloud Storage optimization

Apply now Apply later

* Salary range is an estimate based on our AI, ML, Data Science Salary Index 💰

Job stats:  0  0  0

Tags: Airflow Architecture Azure BigQuery DataOps GCP Google Cloud Kafka Pipelines PySpark Spark Terraform

Regions: Remote/Anywhere Europe
Country: Portugal

More jobs like this