AWS Data Engineer
Aguascalientes, AGU, MX
Capgemini
A global leader in consulting, technology services and digital transformation, we offer an array of integrated services combining technology with deep sector expertise.RH: Raul García
Position: AWS Data Engineer
Location: Aguascalientes, Mexico
Industry - Sector: MALS
What you’ll do?
- Build and maintain end-to-end ETL pipelines on AWS using services like AWS Glue, AWS Lambda, Amazon Kinesis, and Amazon Redshift.
- Implement batch and real-time data processing workflows to integrate data from multiple sources.
- Design and implement data lakes using Amazon S3 to store structured and unstructured data.
- Build and optimize data warehouses using Amazon Redshift, ensuring efficient data storage and retrieval.
- Architect scalable and high-performance data solutions in the AWS Cloud using services such as AWS S3, AWS Glue, Amazon Redshift, Amazon RDS, Amazon Aurora, and Amazon DynamoDB.
- Ensure that data pipelines and data storage are optimized for performance, cost efficiency, and scalability.
- Design and implement data transformation logic using AWS Glue, AWS Lambda, or Apache Spark (running on AWS EMR or Databricks).
- Clean, preprocess, and aggregate data to create datasets suitable for analytics and machine learning.
- Automate the deployment of data pipelines and infrastructure using AWS CloudFormation, AWS CDK, and Terraform.
- Implement CI/CD pipelines for data engineering workflows using AWS CodePipeline, AWS CodeBuild, and AWS CodeDeploy.
- Implement data governance and security best practices, including encryption, access control using AWS IAM, and data masking.
- Ensure compliance with regulatory requirements and internal policies for data security, privacy, and governance.
- Continuously monitor and optimize data pipelines and data storage for performance and cost efficiency.
- Troubleshoot issues related to data processing, pipeline failures, and system performance.
- Collaborate with cross-functional teams (data scientists, analysts, business stakeholders) to understand data needs and provide technical solutions.
- Provide mentorship and guidance to junior data engineers on best practices for AWS data engineering.
- Set up monitoring and logging using AWS CloudWatch to ensure data pipelines are running smoothly.
- Document data engineering processes, workflows, and solutions for internal reference and training.
What you’ll bring:
- Expertise in AWS Glue, Amazon Redshift, Amazon S3, AWS Lambda, Amazon Kinesis, Amazon RDS, AWS Aurora, and AWS DynamoDB.
- Proficient in building and maintaining ETL pipelines using AWS Glue, AWS Lambda, Apache Spark, or AWS Data Pipeline.
- Experience in data extraction, transformation, and loading from diverse sources (relational, NoSQL, flat files, APIs, etc.).
- Experience working with Amazon Redshift and Amazon S3 to build data warehouses and data lakes.
- Strong understanding of data modeling techniques, dimensional modeling, and schema design for analytical purposes.
- Proficiency in Python, SQL, and JavaScript for building data processing scripts, automation tasks, and custom data solutions.
- Experience with Apache Spark, PySpark, and AWS EMR for distributed data processing.
- Hands-on experience with AWS CloudFormation, AWS CDK, Terraform, or other infrastructure-as-code tools for automating deployments.
- Familiarity with CI/CD processes and tools (e.g., AWS CodePipeline, Jenkins, GitLab) to automate data pipeline deployments.
- Expertise in AWS IAM for access control, data encryption (SSE, KMS), and data governance strategies.
- Proficiency in using version control tools like Git to collaborate on development and data engineering projects.
Soft skills:
Work Underpressure, Quality at work, Results Oriented
What can YOU expect in a career with Capgemini?
- Working in a team environment, Consultants will focus on the analysis, design and development of technology-based solutions for Capgemini’s clients.
- You will work alongside technical, functional and industry specialists to assist with the development, implementation and integration of innovative system solutions including methods, techniques and tools.
- You will contribute to client satisfaction by providing timely and responsive value-added services and work products.
- Capgemini offers a competitive compensation and benefits package.
- Headquartered in Paris, France, Capgemini has a presence of more than 340 thousand professionals in Mexico distributed among 3 sites located in Mexico City, Monterrey and Aguascalientes. A deeply multicultural organization.
- Capgemini has developed its own way of working, the Collaborative Business ExperienceTM, and draws on Rightshore, its worldwide delivery model.
You will love this job because
- Capgemini focuses on giving each new hire a YOU-nique experience through our recruitment process and on-boarding program, as well as by helping you to build your own career and professional skills foundation.
- Capgemini provides a collaborative environment that embodies and holds the following stated values close to heart: Honesty, Boldness, Trust, Freedom, Team Spirit, Modesty, and Fun.
- Capgemini cultivates an atmosphere for development that enables YOU to be hands-on, planning for your growth, both horizontally and vertically.
“At Capgemini Mexico, we aim to attract the best talent and are committed to creating a diverse and inclusive work environment, so there is no discrimination based on race, sex, sexual orientation, gender identity or expression, or any other characteristic of a person. All applications welcome and will be considered based on merit against the job and/or experience for the position”
RH: Raul García
Position: AWS Data Engineer
Location: any state of Mexico
Industry - Sector: MALS
What you’ll do?
- Build and maintain end-to-end ETL pipelines on AWS using services like AWS Glue, AWS Lambda, Amazon Kinesis, and Amazon Redshift.
- Implement batch and real-time data processing workflows to integrate data from multiple sources.
- Design and implement data lakes using Amazon S3 to store structured and unstructured data.
- Build and optimize data warehouses using Amazon Redshift, ensuring efficient data storage and retrieval.
- Architect scalable and high-performance data solutions in the AWS Cloud using services such as AWS S3, AWS Glue, Amazon Redshift, Amazon RDS, Amazon Aurora, and Amazon DynamoDB.
- Ensure that data pipelines and data storage are optimized for performance, cost efficiency, and scalability.
- Design and implement data transformation logic using AWS Glue, AWS Lambda, or Apache Spark (running on AWS EMR or Databricks).
- Clean, preprocess, and aggregate data to create datasets suitable for analytics and machine learning.
- Automate the deployment of data pipelines and infrastructure using AWS CloudFormation, AWS CDK, and Terraform.
- Implement CI/CD pipelines for data engineering workflows using AWS CodePipeline, AWS CodeBuild, and AWS CodeDeploy.
- Implement data governance and security best practices, including encryption, access control using AWS IAM, and data masking.
- Ensure compliance with regulatory requirements and internal policies for data security, privacy, and governance.
- Continuously monitor and optimize data pipelines and data storage for performance and cost efficiency.
- Troubleshoot issues related to data processing, pipeline failures, and system performance.
- Collaborate with cross-functional teams (data scientists, analysts, business stakeholders) to understand data needs and provide technical solutions.
- Provide mentorship and guidance to junior data engineers on best practices for AWS data engineering.
- Set up monitoring and logging using AWS CloudWatch to ensure data pipelines are running smoothly.
- Document data engineering processes, workflows, and solutions for internal reference and training.
What you’ll bring:
- Expertise in AWS Glue, Amazon Redshift, Amazon S3, AWS Lambda, Amazon Kinesis, Amazon RDS, AWS Aurora, and AWS DynamoDB.
- Proficient in building and maintaining ETL pipelines using AWS Glue, AWS Lambda, Apache Spark, or AWS Data Pipeline.
- Experience in data extraction, transformation, and loading from diverse sources (relational, NoSQL, flat files, APIs, etc.).
- Experience working with Amazon Redshift and Amazon S3 to build data warehouses and data lakes.
- Strong understanding of data modeling techniques, dimensional modeling, and schema design for analytical purposes.
- Proficiency in Python, SQL, and JavaScript for building data processing scripts, automation tasks, and custom data solutions.
- Experience with Apache Spark, PySpark, and AWS EMR for distributed data processing.
- Hands-on experience with AWS CloudFormation, AWS CDK, Terraform, or other infrastructure-as-code tools for automating deployments.
- Familiarity with CI/CD processes and tools (e.g., AWS CodePipeline, Jenkins, GitLab) to automate data pipeline deployments.
- Expertise in AWS IAM for access control, data encryption (SSE, KMS), and data governance strategies.
- Proficiency in using version control tools like Git to collaborate on development and data engineering projects.
Soft skills:
Work Underpressure, Quality at work, Results Oriented
What can YOU expect in a career with Capgemini?
- Working in a team environment, Consultants will focus on the analysis, design and development of technology-based solutions for Capgemini’s clients.
- You will work alongside technical, functional and industry specialists to assist with the development, implementation and integration of innovative system solutions including methods, techniques and tools.
- You will contribute to client satisfaction by providing timely and responsive value-added services and work products.
- Capgemini offers a competitive compensation and benefits package.
- Headquartered in Paris, France, Capgemini has a presence of more than 340 thousand professionals in Mexico distributed among 3 sites located in Mexico City, Monterrey and Aguascalientes. A deeply multicultural organization.
- Capgemini has developed its own way of working, the Collaborative Business ExperienceTM, and draws on Rightshore, its worldwide delivery model.
You will love this job because
- Capgemini focuses on giving each new hire a YOU-nique experience through our recruitment process and on-boarding program, as well as by helping you to build your own career and professional skills foundation.
- Capgemini provides a collaborative environment that embodies and holds the following stated values close to heart: Honesty, Boldness, Trust, Freedom, Team Spirit, Modesty, and Fun.
- Capgemini cultivates an atmosphere for development that enables YOU to be hands-on, planning for your growth, both horizontally and vertically.
“At Capgemini Mexico, we aim to attract the best talent and are committed to creating a diverse and inclusive work environment, so there is no discrimination based on race, sex, sexual orientation, gender identity or expression, or any other characteristic of a person. All applications welcome and will be considered based on merit against the job and/or experience for the position”
* Salary range is an estimate based on our AI, ML, Data Science Salary Index 💰
Tags: APIs AWS AWS Glue CI/CD CloudFormation Databricks Data governance Data pipelines DynamoDB Engineering ETL Git GitLab JavaScript Jenkins Kinesis Lambda Machine Learning NoSQL Pipelines Privacy PySpark Python Redshift Security Spark SQL Terraform Unstructured data
Perks/benefits: Career development Competitive pay
More jobs like this
Explore more career opportunities
Find even more open roles below ordered by popularity of job title or skills/products/technologies used.