Assistant Manager- Data Engineer

Bangalore, Karnataka, India

KPMG India

Welcome to KPMG International.

View all jobs at KPMG India

Apply now Apply later

 

  1. Responsibilities
  2. An experienced Microsoft Fabric ETL professional capable of integrating into KPMG’s existing processes and workflows.
  3. Collaborating with a global ETL and application development team to share knowledge, best practices, and ensure consistent development standards and to continually enhance development processes based on feedback and emerging industry trends. 
  4. Utilizing modern development tools and frameworks currently in use by the KPMG development team (e.g., Azure Dev Ops).
  5. Ensuring timely reporting and proactive communication with the global development team leadership.
  6. Collaborate with business analysts and business stakeholders to understand data requirements, transformation rules, and relevant accounting business logic.
  7. Design, implement, and optimize ETL processes using Microsoft Fabric, working closely with an established, globally distributed development team. 
  8. Understand the application architecture, tools, and the processes, established and make recommendations to meet needs by application.
  9. Ensure that data is efficiently and accurately processed from source systems into the Microsoft Fabric data layer and transformed across the various levels of transformation inside Fabric.
  10. Develop transformation logic to clean, enrich, aggregate, and otherwise modify data to meet business needs. Implement data validation and error handling procedures to ensure data integrity.
  11. Implement processes to load processed data into Microsoft Fabric data storage or previously worked in technologies like Azure Synapse, Azure Data Lake, etc. Optimize data loading processes to ensure performance and scalability.
  12. Design and deploy ETL pipelines and notebooks using Python and Pyspark within Microsoft Fabric.
  13. Collaborate with data architects to design the data architecture and models that align with the application’s requirements. Implement data schemas, star/snowflake schema designs for data warehouses, and OLAP cubes, as necessary.
  14. Optimize ETL processes for performance and resource utilization. Implement indexing, partitioning, and other performance enhancements to handle large volumes of data efficiently within Fabric.
  15. Develop and execute unit tests, integration tests, and end-to-end tests for ETL processes in Fabric. Troubleshoot and resolve issues, making improvements and adjustments as needed.
  16. Validate data accuracy and consistency throughout the ETL lifecycle.
  17. Create and maintain comprehensive documentation for ETL processes, data sources, data flow diagrams, and data models.
  18. Ensure documentation is up-to-date and accessible to relevant stakeholders.
  19. Ensure all data processing complies with relevant policies and standards.
  20. Required Experience
  21. High proficiency in Microsoft Fabric and related ETL tools (e.g., Azure Data Factory)
  22. Knowledge of database systems (e.g., SQL Server, Azure SQL Database, Synapse Analytics) and understanding of data warehousing concepts and architecture.
  23. Experience with data modeling and schema design.
  24. Familiarity with programming languages used in ETL processes (e.g., Python, Pyspark).
  25. Strong understanding of data engineering principles, including data modeling, data transformation, and data optimization. 
  26. Strong SQL skills for data extraction, transformation, and querying.
  27. Knowledge of accounting principles and logic is highly beneficial.
  28. Working knowledge of US accounting principles and accounting logic would be highly beneficial and desired.

 

  1. Desired Certifications
  2. Microsoft Certified: Fabric Analytics Engineer Associate (strongly preferred)
  3. Microsoft Certified: Azure Data Engineer Associate
  4. Microsoft Certified: Azure Developer Associate
  5. Microsoft Certified: Azure Solutions Architect Expert

 

  1. Responsibilities
  2. An experienced Microsoft Fabric ETL professional capable of integrating into KPMG’s existing processes and workflows.
  3. Collaborating with a global ETL and application development team to share knowledge, best practices, and ensure consistent development standards and to continually enhance development processes based on feedback and emerging industry trends. 
  4. Utilizing modern development tools and frameworks currently in use by the KPMG development team (e.g., Azure Dev Ops).
  5. Ensuring timely reporting and proactive communication with the global development team leadership.
  6. Collaborate with business analysts and business stakeholders to understand data requirements, transformation rules, and relevant accounting business logic.
  7. Design, implement, and optimize ETL processes using Microsoft Fabric, working closely with an established, globally distributed development team. 
  8. Understand the application architecture, tools, and the processes, established and make recommendations to meet needs by application.
  9. Ensure that data is efficiently and accurately processed from source systems into the Microsoft Fabric data layer and transformed across the various levels of transformation inside Fabric.
  10. Develop transformation logic to clean, enrich, aggregate, and otherwise modify data to meet business needs. Implement data validation and error handling procedures to ensure data integrity.
  11. Implement processes to load processed data into Microsoft Fabric data storage or previously worked in technologies like Azure Synapse, Azure Data Lake, etc. Optimize data loading processes to ensure performance and scalability.
  12. Design and deploy ETL pipelines and notebooks using Python and Pyspark within Microsoft Fabric.
  13. Collaborate with data architects to design the data architecture and models that align with the application’s requirements. Implement data schemas, star/snowflake schema designs for data warehouses, and OLAP cubes, as necessary.
  14. Optimize ETL processes for performance and resource utilization. Implement indexing, partitioning, and other performance enhancements to handle large volumes of data efficiently within Fabric.
  15. Develop and execute unit tests, integration tests, and end-to-end tests for ETL processes in Fabric. Troubleshoot and resolve issues, making improvements and adjustments as needed.
  16. Validate data accuracy and consistency throughout the ETL lifecycle.
  17. Create and maintain comprehensive documentation for ETL processes, data sources, data flow diagrams, and data models.
  18. Ensure documentation is up-to-date and accessible to relevant stakeholders.
  19. Ensure all data processing complies with relevant policies and standards.
  20. Required Experience
  21. High proficiency in Microsoft Fabric and related ETL tools (e.g., Azure Data Factory)
  22. Knowledge of database systems (e.g., SQL Server, Azure SQL Database, Synapse Analytics) and understanding of data warehousing concepts and architecture.
  23. Experience with data modeling and schema design.
  24. Familiarity with programming languages used in ETL processes (e.g., Python, Pyspark).
  25. Strong understanding of data engineering principles, including data modeling, data transformation, and data optimization. 
  26. Strong SQL skills for data extraction, transformation, and querying.
  27. Knowledge of accounting principles and logic is highly beneficial.
  28. Working knowledge of US accounting principles and accounting logic would be highly beneficial and desired.

 

  1. Desired Certifications
  2. Microsoft Certified: Fabric Analytics Engineer Associate (strongly preferred)
  3. Microsoft Certified: Azure Data Engineer Associate
  4. Microsoft Certified: Azure Developer Associate
  5. Microsoft Certified: Azure Solutions Architect Expert
Apply now Apply later

* Salary range is an estimate based on our AI, ML, Data Science Salary Index 💰

Job stats:  0  0  0

Tags: Architecture Azure Data Warehousing Engineering ETL OLAP Pipelines PySpark Python Snowflake SQL

Region: Asia/Pacific
Country: India

More jobs like this