Data Engineer - India Data Platforms

MUMBAI GENERAL OFFICE

Applications have closed

Job Location

Mumbai

Job Description

Overview of the job       

Data Engineer– India Data Platforms

This role reports to Director, India Data Platforms, P&G

About India Data Platforms Team:

We take pride in managing the most-valuable asset of company in Digital World, called Data. Our vision is to deliver Data as a competitive advantage for India Business, by building unified data platforms, delivering customized BI tools for managers & empowering insightful business decisions through AI in Data. As a data solutions specialist, you'll be working closely with business stakeholders, collaborating to understand their needs and develop solutions to solve problems in area of supply chain, Sales & Distribution, Consumer Insights & Market performance.

In this role, you'll be constantly learning, staying up to date with industry trends and emerging technologies in data solutions. You'll have the chance to work with a variety of tools and technologies, including big data platforms, machine learning frameworks, and data visualization tools, to build innovative and effective solutions.

So, if you're excited about the possibilities of data, and eager to make a real impact in the world of business, a career in data solutions might be just what you're looking for. Join us and become a part of the future of digital transformation.

Click Here to Hear From the Functional Leader

About P&G IT:

Digital is at the core of P&G’s accelerated growth strategy. With this vision, IT in P&G is deeply embedded into every critical process across business organizations comprising 11+ category units globally creating impactful value through Transformation, Simplification & Innovation. IT in P&G is sub-divided into teams that engage strongly for revolutionizing the business processes to deliver exceptional value & growth - Digital GTM, Digital Manufacturing, Marketing Technologist, Ecommerce, Data Sciences & Analytics, Data Solutions & Engineering, Product Supply.

Responsibilities:

Leading design and development of data and analytics cloud-based platform. Crafting integrated systems, implementing ELT/ ETL jobs to fulfil business deliverables. Performing sophisticated data operations such as data orchestration, transformation, and visualization with large datasets. You will be working with product managers to ensure superior product delivery to drive business value & transformation. Demonstrating standard coding practices to ensure delivery excellence and reusability.

Data Ingestion: Develop and maintain data pipelines to extract data from various sources and load it into Azure and Databricks environments. Data Transformation: Design and implement data transformation processes, including data cleansing, normalization, and aggregation, to ensure data quality and consistency. Data Modeling: Develop and maintain data models and schemas to support efficient data storage and retrieval in Azure and Databricks platforms. Data Warehousing: Design and build data warehouses or data lakes using Azure services such as Azure Data Lake Storage, Databricks DeltaLake Data Integration: Integrate data from multiple sources, both on-premises and cloud-based, using Azure Data Factory or other relevant tools. Data Governance: Implement data governance practices, including data security, privacy, and compliance, to ensure data integrity and regulatory compliance. Performance Optimization: Optimize data pipelines and queries for improved performance and scalability in Azure and Databricks environments. Monitoring and Troubleshooting: Monitor data pipelines, identify and resolve performance issues, and troubleshoot data-related problems in collaboration with other teams. Data Visualization: Build BI reports to enable faster decision making. Collaboration: Work with product managers to ensure superior product delivery to drive business value & transformation Documentation: Document data engineering processes, data flows, and system configurations for future reference and knowledge sharing.

Qualifications:

Experience: Bachelor's or master's degree in computer science, data engineering, or a related field, along with 2+ year work experience in data engineering and cloud platforms. Azure and Databricks: Strong proficiency in Azure services such as Azure Data Factory, Azure Databricks, Azure SQL Database, Azure Data Lake Storage, and Azure Synapse Analytics. ETL Tools: Experience with ETL (Extract, Transform, Load) tools and frameworks, such as Apache Spark, Databricks Delta, or Azure Data Factory, for data integration and transformation. Programming: Proficiency in programming languages such as PySpark, Python, SQL, or Scala for data manipulation, scripting, and automation. Data Modeling: Knowledge of data modeling techniques and experience with data modeling tools. Database Technologies: Familiarity with relational databases (e.g., SQL Server) for data storage and retrieval. Data Warehousing: Understanding of data warehousing concepts, dimensional modeling, and experience with data warehousing technologies such as Azure Synapse Analytics or Azure SQL Data Warehouse or Azure Databricks DeltaLake Data Governance: Knowledge of data governance principles, data security, privacy regulations (e.g., GDPR, CCPA), and experience implementing data governance practices. Data Visualization: Experience of working with Microsoft Power BI to build semantic data model & BI reports/dashboards. Cloud Computing: Familiarity with cloud computing concepts and experience working with cloud platforms, particularly Microsoft Azure. Problem-Solving: Strong analytical and problem-solving skills to identify and resolve data-related issues. Proficiency in DevOps Tools and CICD tools (e.g. Azure DevOps, Chef, Puppet, Github)

About us

We produce globally recognized brands, and we grow the best business leaders in the industry. With a portfolio of trusted brands as diverse as ours, it is paramount our leaders are able to lead with courage the vast array of brands, categories and functions. We serve consumers around the world with one of the strongest portfolios of trusted, quality, leadership brands, including Always®, Ariel®, Gillette®, Head & Shoulders®, Herbal Essences®, Oral-B®, Pampers®, Pantene®, Tampax® and more. Our community includes operations in approximately 70 countries worldwide.

Visit http://www.pg.com to know more.

We are an equal opportunity employer and value diversity at our company. We do not discriminate against individuals on the basis of race, color, gender, age, national origin, religion, sexual orientation, gender identity or expression, marital status, citizenship, disability, HIV/AIDS status, or any other legally protected factor.

Job Qualifications

  • Qualifications:

  • Experience: Bachelor's or master's degree in computer science, data engineering, or a related field, along with 2+ year work experience in data engineering and cloud platforms.
  • Azure and Databricks: Strong proficiency in Azure services such as Azure Data Factory, Azure Databricks, Azure SQL Database, Azure Data Lake Storage, and Azure Synapse Analytics.
  • ETL Tools: Experience with ETL (Extract, Transform, Load) tools and frameworks, such as Apache Spark, Databricks Delta, or Azure Data Factory, for data integration and transformation.
  • Programming: Proficiency in programming languages such as PySpark, Python, SQL, or Scala for data manipulation, scripting, and automation.
  • Data Modeling: Knowledge of data modeling techniques and experience with data modeling tools.
  • Database Technologies: Familiarity with relational databases (e.g., SQL Server) for data storage and retrieval.
  • Data Warehousing: Understanding of data warehousing concepts, dimensional modeling, and experience with data warehousing technologies such as Azure Synapse Analytics or Azure SQL Data Warehouse or Azure Databricks DeltaLake
  • Data Governance: Knowledge of data governance principles, data security, privacy regulations (e.g., GDPR, CCPA), and experience implementing data governance practices.
  • Data Visualization: Experience of working with Microsoft Power BI to build semantic data model & BI reports/dashboards.
  • Cloud Computing: Familiarity with cloud computing concepts and experience working with cloud platforms, particularly Microsoft Azure.
  • Problem-Solving: Strong analytical and problem-solving skills to identify and resolve data-related issues.
  • Proficiency in DevOps Tools and CICD tools (e.g. Azure DevOps, Chef, Puppet, Github)

Job Schedule

Full time

Job Number

R000114602

Job Segmentation

Experienced Professionals (Job Segmentation)

* Salary range is an estimate based on our AI, ML, Data Science Salary Index 💰

Job stats:  7  2  0
Category: Engineering Jobs

Tags: Azure Big Data Computer Science Databricks Data governance DataOps Data pipelines Data quality Data visualization Data warehouse Data Warehousing DevOps E-commerce ELT Engineering ETL GitHub Machine Learning Pipelines Power BI Privacy Puppet PySpark Python RDBMS Scala Security Spark SQL

Perks/benefits: Career development

Region: Asia/Pacific
Country: India

More jobs like this