Connecting people I'd hire with companies I'd work at

Matt Wallaert
35
companies
9,198
Jobs

AWS Data Engineer

Capgemini

Capgemini

Data Science
Mexico
Posted on Dec 18, 2024

RH: Raul García

Position: AWS Data Engineer

Location: Aguascalientes, Mexico

Industry - Sector: MALS

What you’ll do?

  • Build and maintain end-to-end ETL pipelines on AWS using services like AWS Glue, AWS Lambda, Amazon Kinesis, and Amazon Redshift.
  • Implement batch and real-time data processing workflows to integrate data from multiple sources.
  • Design and implement data lakes using Amazon S3 to store structured and unstructured data.
  • Build and optimize data warehouses using Amazon Redshift, ensuring efficient data storage and retrieval.
  • Architect scalable and high-performance data solutions in the AWS Cloud using services such as AWS S3, AWS Glue, Amazon Redshift, Amazon RDS, Amazon Aurora, and Amazon DynamoDB.
  • Ensure that data pipelines and data storage are optimized for performance, cost efficiency, and scalability.
  • Design and implement data transformation logic using AWS Glue, AWS Lambda, or Apache Spark (running on AWS EMR or Databricks).
  • Clean, preprocess, and aggregate data to create datasets suitable for analytics and machine learning.
  • Automate the deployment of data pipelines and infrastructure using AWS CloudFormation, AWS CDK, and Terraform.
  • Implement CI/CD pipelines for data engineering workflows using AWS CodePipeline, AWS CodeBuild, and AWS CodeDeploy.
  • Implement data governance and security best practices, including encryption, access control using AWS IAM, and data masking.
  • Ensure compliance with regulatory requirements and internal policies for data security, privacy, and governance.
  • Continuously monitor and optimize data pipelines and data storage for performance and cost efficiency.
  • Troubleshoot issues related to data processing, pipeline failures, and system performance.
  • Collaborate with cross-functional teams (data scientists, analysts, business stakeholders) to understand data needs and provide technical solutions.
  • Provide mentorship and guidance to junior data engineers on best practices for AWS data engineering.
  • Set up monitoring and logging using AWS CloudWatch to ensure data pipelines are running smoothly.
  • Document data engineering processes, workflows, and solutions for internal reference and training.

What you’ll bring:

  • Expertise in AWS Glue, Amazon Redshift, Amazon S3, AWS Lambda, Amazon Kinesis, Amazon RDS, AWS Aurora, and AWS DynamoDB.
  • Proficient in building and maintaining ETL pipelines using AWS Glue, AWS Lambda, Apache Spark, or AWS Data Pipeline.
  • Experience in data extraction, transformation, and loading from diverse sources (relational, NoSQL, flat files, APIs, etc.).
  • Experience working with Amazon Redshift and Amazon S3 to build data warehouses and data lakes.
  • Strong understanding of data modeling techniques, dimensional modeling, and schema design for analytical purposes.
  • Proficiency in Python, SQL, and JavaScript for building data processing scripts, automation tasks, and custom data solutions.
  • Experience with Apache Spark, PySpark, and AWS EMR for distributed data processing.
  • Hands-on experience with AWS CloudFormation, AWS CDK, Terraform, or other infrastructure-as-code tools for automating deployments.
  • Familiarity with CI/CD processes and tools (e.g., AWS CodePipeline, Jenkins, GitLab) to automate data pipeline deployments.
  • Expertise in AWS IAM for access control, data encryption (SSE, KMS), and data governance strategies.
  • Proficiency in using version control tools like Git to collaborate on development and data engineering projects.

Soft skills:

Work Underpressure, Quality at work, Results Oriented

What can YOU expect in a career with Capgemini?

  • Working in a team environment, Consultants will focus on the analysis, design and development of technology-based solutions for Capgemini’s clients.
  • You will work alongside technical, functional and industry specialists to assist with the development, implementation and integration of innovative system solutions including methods, techniques and tools.
  • You will contribute to client satisfaction by providing timely and responsive value-added services and work products.
  • Capgemini offers a competitive compensation and benefits package.
  • Headquartered in Paris, France, Capgemini has a presence of more than 340 thousand professionals in Mexico distributed among 3 sites located in Mexico City, Monterrey and Aguascalientes. A deeply multicultural organization.
  • Capgemini has developed its own way of working, the Collaborative Business ExperienceTM, and draws on Rightshore, its worldwide delivery model.

You will love this job because

  • Capgemini focuses on giving each new hire a YOU-nique experience through our recruitment process and on-boarding program, as well as by helping you to build your own career and professional skills foundation.
  • Capgemini provides a collaborative environment that embodies and holds the following stated values close to heart: Honesty, Boldness, Trust, Freedom, Team Spirit, Modesty, and Fun.
  • Capgemini cultivates an atmosphere for development that enables YOU to be hands-on, planning for your growth, both horizontally and vertically.

At Capgemini Mexico, we aim to attract the best talent and are committed to creating a diverse and inclusive work environment, so there is no discrimination based on race, sex, sexual orientation, gender identity or expression, or any other characteristic of a person. All applications welcome and will be considered based on merit against the job and/or experience for the position”

RH: Raul García

Position: AWS Data Engineer

Location: any state of Mexico

Industry - Sector: MALS

What you’ll do?

  • Build and maintain end-to-end ETL pipelines on AWS using services like AWS Glue, AWS Lambda, Amazon Kinesis, and Amazon Redshift.
  • Implement batch and real-time data processing workflows to integrate data from multiple sources.
  • Design and implement data lakes using Amazon S3 to store structured and unstructured data.
  • Build and optimize data warehouses using Amazon Redshift, ensuring efficient data storage and retrieval.
  • Architect scalable and high-performance data solutions in the AWS Cloud using services such as AWS S3, AWS Glue, Amazon Redshift, Amazon RDS, Amazon Aurora, and Amazon DynamoDB.
  • Ensure that data pipelines and data storage are optimized for performance, cost efficiency, and scalability.
  • Design and implement data transformation logic using AWS Glue, AWS Lambda, or Apache Spark (running on AWS EMR or Databricks).
  • Clean, preprocess, and aggregate data to create datasets suitable for analytics and machine learning.
  • Automate the deployment of data pipelines and infrastructure using AWS CloudFormation, AWS CDK, and Terraform.
  • Implement CI/CD pipelines for data engineering workflows using AWS CodePipeline, AWS CodeBuild, and AWS CodeDeploy.
  • Implement data governance and security best practices, including encryption, access control using AWS IAM, and data masking.
  • Ensure compliance with regulatory requirements and internal policies for data security, privacy, and governance.
  • Continuously monitor and optimize data pipelines and data storage for performance and cost efficiency.
  • Troubleshoot issues related to data processing, pipeline failures, and system performance.
  • Collaborate with cross-functional teams (data scientists, analysts, business stakeholders) to understand data needs and provide technical solutions.
  • Provide mentorship and guidance to junior data engineers on best practices for AWS data engineering.
  • Set up monitoring and logging using AWS CloudWatch to ensure data pipelines are running smoothly.
  • Document data engineering processes, workflows, and solutions for internal reference and training.

What you’ll bring:

  • Expertise in AWS Glue, Amazon Redshift, Amazon S3, AWS Lambda, Amazon Kinesis, Amazon RDS, AWS Aurora, and AWS DynamoDB.
  • Proficient in building and maintaining ETL pipelines using AWS Glue, AWS Lambda, Apache Spark, or AWS Data Pipeline.
  • Experience in data extraction, transformation, and loading from diverse sources (relational, NoSQL, flat files, APIs, etc.).
  • Experience working with Amazon Redshift and Amazon S3 to build data warehouses and data lakes.
  • Strong understanding of data modeling techniques, dimensional modeling, and schema design for analytical purposes.
  • Proficiency in Python, SQL, and JavaScript for building data processing scripts, automation tasks, and custom data solutions.
  • Experience with Apache Spark, PySpark, and AWS EMR for distributed data processing.
  • Hands-on experience with AWS CloudFormation, AWS CDK, Terraform, or other infrastructure-as-code tools for automating deployments.
  • Familiarity with CI/CD processes and tools (e.g., AWS CodePipeline, Jenkins, GitLab) to automate data pipeline deployments.
  • Expertise in AWS IAM for access control, data encryption (SSE, KMS), and data governance strategies.
  • Proficiency in using version control tools like Git to collaborate on development and data engineering projects.

Soft skills:

Work Underpressure, Quality at work, Results Oriented

What can YOU expect in a career with Capgemini?

  • Working in a team environment, Consultants will focus on the analysis, design and development of technology-based solutions for Capgemini’s clients.
  • You will work alongside technical, functional and industry specialists to assist with the development, implementation and integration of innovative system solutions including methods, techniques and tools.
  • You will contribute to client satisfaction by providing timely and responsive value-added services and work products.
  • Capgemini offers a competitive compensation and benefits package.
  • Headquartered in Paris, France, Capgemini has a presence of more than 340 thousand professionals in Mexico distributed among 3 sites located in Mexico City, Monterrey and Aguascalientes. A deeply multicultural organization.
  • Capgemini has developed its own way of working, the Collaborative Business ExperienceTM, and draws on Rightshore, its worldwide delivery model.

You will love this job because

  • Capgemini focuses on giving each new hire a YOU-nique experience through our recruitment process and on-boarding program, as well as by helping you to build your own career and professional skills foundation.
  • Capgemini provides a collaborative environment that embodies and holds the following stated values close to heart: Honesty, Boldness, Trust, Freedom, Team Spirit, Modesty, and Fun.
  • Capgemini cultivates an atmosphere for development that enables YOU to be hands-on, planning for your growth, both horizontally and vertically.

At Capgemini Mexico, we aim to attract the best talent and are committed to creating a diverse and inclusive work environment, so there is no discrimination based on race, sex, sexual orientation, gender identity or expression, or any other characteristic of a person. All applications welcome and will be considered based on merit against the job and/or experience for the position”