Data Engineer (Remote)
Remote, Guadalajara
As a Data Engineer, you will play a critical role in building and optimizing data infrastructure to support analytics and machine learning initiatives. You will design scalable data pipelines, develop ETL workflows, and ensure the efficient integration of data across various platforms using industry-standard technologies and tools. Your work will enable cross-functional teams to derive insights and make data-driven decisions.
Responsibilities:
- Design, build, and maintain scalable data pipelines and workflows using multiple Cloud services and tools
- Collaborate with data scientists, machine learning engineers, and business stakeholders to understand data requirements and deliver appropriate solutions.
- Optimize data storage solutions and implement best practices for data governance, security, and performance.
- Implement Python-based solutions for data processing and analysis.
- Build and refine CI/CD processes to improve data workflows and ensure seamless deployments.
- Monitor and troubleshoot data pipelines to ensure reliability and minimize downtime.
- Stay up to date with advancements in data engineering and cloud computing.
Qualifications and Requirements:
- Bachelor's or Master's degree in Data Engineering, Computer Science or a related field (or equivalent practical experience).
- 5+ years of experience in designing and implementing data pipelines and ETL workflows.
- Proven track record of delivering production-grade software.
- Strong problem-solving skills and attention to detail.
- Excellent English communication.
- Strong problem-solving and analytical skills, with the ability to work collaboratively in a team environment.
- Ability to thrive in a fast-paced, agile environment, with the capability to drive frontend and backend architectural decisions.
- Strong experience with Python programming language.
- Exceptional SQL proficiency for querying and managing complex data structures.
- Deep experience with Databricks for managing and optimizing large-scale data systems.
- Strong experience with at least one cloud service (AWS, Azure, or GCP)
- Deep understanding of database systems, data warehousing, and data modeling techniques.
- Familiarity with distributed computing and Big Data frameworks like Apache Spark or Hadoop.
Preferred Skills:
- Experience in the transportation and logistics industry.
- Knowledge of IoT data integration and real-time data processing.
- Familiarity with MLOps workflows and supporting machine learning teams..
- Experience with data visualization tools like Power BI or Tableau.
* Salary range is an estimate based on our AI, ML, Data Science Salary Index 💰
Tags: Agile AWS Azure Big Data CI/CD Computer Science Databricks Data governance Data pipelines Data visualization Data Warehousing Engineering ETL GCP Hadoop Machine Learning MLOps Pipelines Power BI Python Security Spark SQL Tableau
Perks/benefits: Startup environment
More jobs like this
Explore more career opportunities
Find even more open roles below ordered by popularity of job title or skills/products/technologies used.