AWS Data Engineer

Aguascalientes, AGU, MX

Capgemini

A global leader in consulting, technology services and digital transformation, we offer an array of integrated services combining technology with deep sector expertise.

View all jobs at Capgemini

Apply now Apply later

RH: Raul García

Position: AWS Data Engineer

Location: Aguascalientes, Mexico

Industry - Sector: MALS

What you’ll do?

  • Build and maintain end-to-end ETL pipelines on AWS using services like AWS Glue, AWS Lambda, Amazon Kinesis, and Amazon Redshift.
  • Implement batch and real-time data processing workflows to integrate data from multiple sources.
  • Design and implement data lakes using Amazon S3 to store structured and unstructured data.
  • Build and optimize data warehouses using Amazon Redshift, ensuring efficient data storage and retrieval.
  • Architect scalable and high-performance data solutions in the AWS Cloud using services such as AWS S3, AWS Glue, Amazon Redshift, Amazon RDS, Amazon Aurora, and Amazon DynamoDB.
  • Ensure that data pipelines and data storage are optimized for performance, cost efficiency, and scalability.
  • Design and implement data transformation logic using AWS Glue, AWS Lambda, or Apache Spark (running on AWS EMR or Databricks).
  • Clean, preprocess, and aggregate data to create datasets suitable for analytics and machine learning.
  • Automate the deployment of data pipelines and infrastructure using AWS CloudFormation, AWS CDK, and Terraform.
  • Implement CI/CD pipelines for data engineering workflows using AWS CodePipeline, AWS CodeBuild, and AWS CodeDeploy.
  • Implement data governance and security best practices, including encryption, access control using AWS IAM, and data masking.
  • Ensure compliance with regulatory requirements and internal policies for data security, privacy, and governance.
  • Continuously monitor and optimize data pipelines and data storage for performance and cost efficiency.
  • Troubleshoot issues related to data processing, pipeline failures, and system performance.
  • Collaborate with cross-functional teams (data scientists, analysts, business stakeholders) to understand data needs and provide technical solutions.
  • Provide mentorship and guidance to junior data engineers on best practices for AWS data engineering.
  • Set up monitoring and logging using AWS CloudWatch to ensure data pipelines are running smoothly.
  • Document data engineering processes, workflows, and solutions for internal reference and training.

What you’ll bring:

  • Expertise in AWS Glue, Amazon Redshift, Amazon S3, AWS Lambda, Amazon Kinesis, Amazon RDS, AWS Aurora, and AWS DynamoDB.
  • Proficient in building and maintaining ETL pipelines using AWS Glue, AWS Lambda, Apache Spark, or AWS Data Pipeline.
  • Experience in data extraction, transformation, and loading from diverse sources (relational, NoSQL, flat files, APIs, etc.).
  • Experience working with Amazon Redshift and Amazon S3 to build data warehouses and data lakes.
  • Strong understanding of data modeling techniques, dimensional modeling, and schema design for analytical purposes.
  • Proficiency in Python, SQL, and JavaScript for building data processing scripts, automation tasks, and custom data solutions.
  • Experience with Apache Spark, PySpark, and AWS EMR for distributed data processing.
  • Hands-on experience with AWS CloudFormation, AWS CDK, Terraform, or other infrastructure-as-code tools for automating deployments.
  • Familiarity with CI/CD processes and tools (e.g., AWS CodePipeline, Jenkins, GitLab) to automate data pipeline deployments.
  • Expertise in AWS IAM for access control, data encryption (SSE, KMS), and data governance strategies.
  • Proficiency in using version control tools like Git to collaborate on development and data engineering projects.

Soft skills:

Work Underpressure, Quality at work, Results Oriented

What can YOU expect in a career with Capgemini?

  • Working in a team environment, Consultants will focus on the analysis, design and development of technology-based solutions for Capgemini’s clients.
  • You will work alongside technical, functional and industry specialists to assist with the development, implementation and integration of innovative system solutions including methods, techniques and tools.
  • You will contribute to client satisfaction by providing timely and responsive value-added services and work products.
  • Capgemini offers a competitive compensation and benefits package.
  • Headquartered in Paris, France, Capgemini has a presence of more than 340 thousand professionals in Mexico distributed among 3 sites located in Mexico City, Monterrey and Aguascalientes. A deeply multicultural organization.
  • Capgemini has developed its own way of working, the Collaborative Business ExperienceTM, and draws on Rightshore, its worldwide delivery model.

You will love this job because

  • Capgemini focuses on giving each new hire a YOU-nique experience through our recruitment process and on-boarding program, as well as by helping you to build your own career and professional skills foundation.
  • Capgemini provides a collaborative environment that embodies and holds the following stated values close to heart: Honesty, Boldness, Trust, Freedom, Team Spirit, Modesty, and Fun.
  • Capgemini cultivates an atmosphere for development that enables YOU to be hands-on, planning for your growth, both horizontally and vertically.

At Capgemini Mexico, we aim to attract the best talent and are committed to creating a diverse and inclusive work environment, so there is no discrimination based on race, sex, sexual orientation, gender identity or expression, or any other characteristic of a person. All applications welcome and will be considered based on merit against the job and/or experience for the position”

RH: Raul García

Position: AWS Data Engineer

Location: any state of Mexico

Industry - Sector: MALS

What you’ll do?

  • Build and maintain end-to-end ETL pipelines on AWS using services like AWS Glue, AWS Lambda, Amazon Kinesis, and Amazon Redshift.
  • Implement batch and real-time data processing workflows to integrate data from multiple sources.
  • Design and implement data lakes using Amazon S3 to store structured and unstructured data.
  • Build and optimize data warehouses using Amazon Redshift, ensuring efficient data storage and retrieval.
  • Architect scalable and high-performance data solutions in the AWS Cloud using services such as AWS S3, AWS Glue, Amazon Redshift, Amazon RDS, Amazon Aurora, and Amazon DynamoDB.
  • Ensure that data pipelines and data storage are optimized for performance, cost efficiency, and scalability.
  • Design and implement data transformation logic using AWS Glue, AWS Lambda, or Apache Spark (running on AWS EMR or Databricks).
  • Clean, preprocess, and aggregate data to create datasets suitable for analytics and machine learning.
  • Automate the deployment of data pipelines and infrastructure using AWS CloudFormation, AWS CDK, and Terraform.
  • Implement CI/CD pipelines for data engineering workflows using AWS CodePipeline, AWS CodeBuild, and AWS CodeDeploy.
  • Implement data governance and security best practices, including encryption, access control using AWS IAM, and data masking.
  • Ensure compliance with regulatory requirements and internal policies for data security, privacy, and governance.
  • Continuously monitor and optimize data pipelines and data storage for performance and cost efficiency.
  • Troubleshoot issues related to data processing, pipeline failures, and system performance.
  • Collaborate with cross-functional teams (data scientists, analysts, business stakeholders) to understand data needs and provide technical solutions.
  • Provide mentorship and guidance to junior data engineers on best practices for AWS data engineering.
  • Set up monitoring and logging using AWS CloudWatch to ensure data pipelines are running smoothly.
  • Document data engineering processes, workflows, and solutions for internal reference and training.

What you’ll bring:

  • Expertise in AWS Glue, Amazon Redshift, Amazon S3, AWS Lambda, Amazon Kinesis, Amazon RDS, AWS Aurora, and AWS DynamoDB.
  • Proficient in building and maintaining ETL pipelines using AWS Glue, AWS Lambda, Apache Spark, or AWS Data Pipeline.
  • Experience in data extraction, transformation, and loading from diverse sources (relational, NoSQL, flat files, APIs, etc.).
  • Experience working with Amazon Redshift and Amazon S3 to build data warehouses and data lakes.
  • Strong understanding of data modeling techniques, dimensional modeling, and schema design for analytical purposes.
  • Proficiency in Python, SQL, and JavaScript for building data processing scripts, automation tasks, and custom data solutions.
  • Experience with Apache Spark, PySpark, and AWS EMR for distributed data processing.
  • Hands-on experience with AWS CloudFormation, AWS CDK, Terraform, or other infrastructure-as-code tools for automating deployments.
  • Familiarity with CI/CD processes and tools (e.g., AWS CodePipeline, Jenkins, GitLab) to automate data pipeline deployments.
  • Expertise in AWS IAM for access control, data encryption (SSE, KMS), and data governance strategies.
  • Proficiency in using version control tools like Git to collaborate on development and data engineering projects.

Soft skills:

Work Underpressure, Quality at work, Results Oriented

What can YOU expect in a career with Capgemini?

  • Working in a team environment, Consultants will focus on the analysis, design and development of technology-based solutions for Capgemini’s clients.
  • You will work alongside technical, functional and industry specialists to assist with the development, implementation and integration of innovative system solutions including methods, techniques and tools.
  • You will contribute to client satisfaction by providing timely and responsive value-added services and work products.
  • Capgemini offers a competitive compensation and benefits package.
  • Headquartered in Paris, France, Capgemini has a presence of more than 340 thousand professionals in Mexico distributed among 3 sites located in Mexico City, Monterrey and Aguascalientes. A deeply multicultural organization.
  • Capgemini has developed its own way of working, the Collaborative Business ExperienceTM, and draws on Rightshore, its worldwide delivery model.

You will love this job because

  • Capgemini focuses on giving each new hire a YOU-nique experience through our recruitment process and on-boarding program, as well as by helping you to build your own career and professional skills foundation.
  • Capgemini provides a collaborative environment that embodies and holds the following stated values close to heart: Honesty, Boldness, Trust, Freedom, Team Spirit, Modesty, and Fun.
  • Capgemini cultivates an atmosphere for development that enables YOU to be hands-on, planning for your growth, both horizontally and vertically.

At Capgemini Mexico, we aim to attract the best talent and are committed to creating a diverse and inclusive work environment, so there is no discrimination based on race, sex, sexual orientation, gender identity or expression, or any other characteristic of a person. All applications welcome and will be considered based on merit against the job and/or experience for the position”

Apply now Apply later

* Salary range is an estimate based on our AI, ML, Data Science Salary Index 💰

Job stats:  0  0  0
Category: Engineering Jobs

Tags: APIs AWS AWS Glue CI/CD CloudFormation Databricks Data governance Data pipelines DynamoDB Engineering ETL Git GitLab JavaScript Jenkins Kinesis Lambda Machine Learning NoSQL Pipelines Privacy PySpark Python Redshift Security Spark SQL Terraform Unstructured data

Perks/benefits: Career development Competitive pay

Region: North America
Country: Mexico

More jobs like this