Assistant Manager- Data Engineer
Bangalore, Karnataka, India
- Responsibilities
- An experienced Microsoft Fabric ETL professional capable of integrating into KPMG’s existing processes and workflows.
- Collaborating with a global ETL and application development team to share knowledge, best practices, and ensure consistent development standards and to continually enhance development processes based on feedback and emerging industry trends.
- Utilizing modern development tools and frameworks currently in use by the KPMG development team (e.g., Azure Dev Ops).
- Ensuring timely reporting and proactive communication with the global development team leadership.
- Collaborate with business analysts and business stakeholders to understand data requirements, transformation rules, and relevant accounting business logic.
- Design, implement, and optimize ETL processes using Microsoft Fabric, working closely with an established, globally distributed development team.
- Understand the application architecture, tools, and the processes, established and make recommendations to meet needs by application.
- Ensure that data is efficiently and accurately processed from source systems into the Microsoft Fabric data layer and transformed across the various levels of transformation inside Fabric.
- Develop transformation logic to clean, enrich, aggregate, and otherwise modify data to meet business needs. Implement data validation and error handling procedures to ensure data integrity.
- Implement processes to load processed data into Microsoft Fabric data storage or previously worked in technologies like Azure Synapse, Azure Data Lake, etc. Optimize data loading processes to ensure performance and scalability.
- Design and deploy ETL pipelines and notebooks using Python and Pyspark within Microsoft Fabric.
- Collaborate with data architects to design the data architecture and models that align with the application’s requirements. Implement data schemas, star/snowflake schema designs for data warehouses, and OLAP cubes, as necessary.
- Optimize ETL processes for performance and resource utilization. Implement indexing, partitioning, and other performance enhancements to handle large volumes of data efficiently within Fabric.
- Develop and execute unit tests, integration tests, and end-to-end tests for ETL processes in Fabric. Troubleshoot and resolve issues, making improvements and adjustments as needed.
- Validate data accuracy and consistency throughout the ETL lifecycle.
- Create and maintain comprehensive documentation for ETL processes, data sources, data flow diagrams, and data models.
- Ensure documentation is up-to-date and accessible to relevant stakeholders.
- Ensure all data processing complies with relevant policies and standards.
- Required Experience
- High proficiency in Microsoft Fabric and related ETL tools (e.g., Azure Data Factory)
- Knowledge of database systems (e.g., SQL Server, Azure SQL Database, Synapse Analytics) and understanding of data warehousing concepts and architecture.
- Experience with data modeling and schema design.
- Familiarity with programming languages used in ETL processes (e.g., Python, Pyspark).
- Strong understanding of data engineering principles, including data modeling, data transformation, and data optimization.
- Strong SQL skills for data extraction, transformation, and querying.
- Knowledge of accounting principles and logic is highly beneficial.
- Working knowledge of US accounting principles and accounting logic would be highly beneficial and desired.
- Desired Certifications
- Microsoft Certified: Fabric Analytics Engineer Associate (strongly preferred)
- Microsoft Certified: Azure Data Engineer Associate
- Microsoft Certified: Azure Developer Associate
- Microsoft Certified: Azure Solutions Architect Expert
- Responsibilities
- An experienced Microsoft Fabric ETL professional capable of integrating into KPMG’s existing processes and workflows.
- Collaborating with a global ETL and application development team to share knowledge, best practices, and ensure consistent development standards and to continually enhance development processes based on feedback and emerging industry trends.
- Utilizing modern development tools and frameworks currently in use by the KPMG development team (e.g., Azure Dev Ops).
- Ensuring timely reporting and proactive communication with the global development team leadership.
- Collaborate with business analysts and business stakeholders to understand data requirements, transformation rules, and relevant accounting business logic.
- Design, implement, and optimize ETL processes using Microsoft Fabric, working closely with an established, globally distributed development team.
- Understand the application architecture, tools, and the processes, established and make recommendations to meet needs by application.
- Ensure that data is efficiently and accurately processed from source systems into the Microsoft Fabric data layer and transformed across the various levels of transformation inside Fabric.
- Develop transformation logic to clean, enrich, aggregate, and otherwise modify data to meet business needs. Implement data validation and error handling procedures to ensure data integrity.
- Implement processes to load processed data into Microsoft Fabric data storage or previously worked in technologies like Azure Synapse, Azure Data Lake, etc. Optimize data loading processes to ensure performance and scalability.
- Design and deploy ETL pipelines and notebooks using Python and Pyspark within Microsoft Fabric.
- Collaborate with data architects to design the data architecture and models that align with the application’s requirements. Implement data schemas, star/snowflake schema designs for data warehouses, and OLAP cubes, as necessary.
- Optimize ETL processes for performance and resource utilization. Implement indexing, partitioning, and other performance enhancements to handle large volumes of data efficiently within Fabric.
- Develop and execute unit tests, integration tests, and end-to-end tests for ETL processes in Fabric. Troubleshoot and resolve issues, making improvements and adjustments as needed.
- Validate data accuracy and consistency throughout the ETL lifecycle.
- Create and maintain comprehensive documentation for ETL processes, data sources, data flow diagrams, and data models.
- Ensure documentation is up-to-date and accessible to relevant stakeholders.
- Ensure all data processing complies with relevant policies and standards.
- Required Experience
- High proficiency in Microsoft Fabric and related ETL tools (e.g., Azure Data Factory)
- Knowledge of database systems (e.g., SQL Server, Azure SQL Database, Synapse Analytics) and understanding of data warehousing concepts and architecture.
- Experience with data modeling and schema design.
- Familiarity with programming languages used in ETL processes (e.g., Python, Pyspark).
- Strong understanding of data engineering principles, including data modeling, data transformation, and data optimization.
- Strong SQL skills for data extraction, transformation, and querying.
- Knowledge of accounting principles and logic is highly beneficial.
- Working knowledge of US accounting principles and accounting logic would be highly beneficial and desired.
- Desired Certifications
- Microsoft Certified: Fabric Analytics Engineer Associate (strongly preferred)
- Microsoft Certified: Azure Data Engineer Associate
- Microsoft Certified: Azure Developer Associate
- Microsoft Certified: Azure Solutions Architect Expert
* Salary range is an estimate based on our AI, ML, Data Science Salary Index 💰
Job stats:
0
0
0
Categories:
Engineering Jobs
Leadership Jobs
Tags: Architecture Azure Data Warehousing Engineering ETL OLAP Pipelines PySpark Python Snowflake SQL
Region:
Asia/Pacific
Country:
India
More jobs like this
Explore more career opportunities
Find even more open roles below ordered by popularity of job title or skills/products/technologies used.
Staff Machine Learning Engineer jobsData Engineer II jobsStaff Data Scientist jobsBI Developer jobsData Scientist II jobsPrincipal Data Engineer jobsData Manager jobsJunior Data Analyst jobsData Science Manager jobsSenior AI Engineer jobsResearch Scientist jobsBusiness Data Analyst jobsData Specialist jobsPrincipal Software Engineer jobsData Science Intern jobsLead Data Analyst jobsData Analyst Intern jobsData Analyst II jobsSr. Data Scientist jobsBI Analyst jobsData Engineer III jobsSoftware Engineer II jobsSoftware Engineer, Machine Learning jobsAzure Data Engineer jobsJunior Data Engineer jobs
Snowflake jobsLinux jobsEconomics jobsOpen Source jobsComputer Vision jobsRDBMS jobsData Warehousing jobsHadoop jobsGoogle Cloud jobsMLOps jobsAirflow jobsKafka jobsNoSQL jobsClassification jobsBanking jobsJavaScript jobsScala jobsKPIs jobsData warehouse jobsStreaming jobsScikit-learn jobsPhysics jobsLooker jobsOracle jobsR&D jobs
PostgreSQL jobsPySpark jobsTerraform jobsGitHub jobsCX jobsRobotics jobsPandas jobsSAS jobsBigQuery jobsScrum jobsData Mining jobsIndustrial jobsDistributed Systems jobsRedshift jobsJira jobsUnstructured data jobsPharma jobsMicroservices jobsdbt jobsJenkins jobsGPT jobsRAG jobsE-commerce jobsReact jobsMySQL jobs