Assistant Manager- Data Engineer
Bangalore, Karnataka, India
- Responsibilities
- An experienced Microsoft Fabric ETL professional capable of integrating into KPMG’s existing processes and workflows.
- Collaborating with a global ETL and application development team to share knowledge, best practices, and ensure consistent development standards and to continually enhance development processes based on feedback and emerging industry trends.
- Utilizing modern development tools and frameworks currently in use by the KPMG development team (e.g., Azure Dev Ops).
- Ensuring timely reporting and proactive communication with the global development team leadership.
- Collaborate with business analysts and business stakeholders to understand data requirements, transformation rules, and relevant accounting business logic.
- Design, implement, and optimize ETL processes using Microsoft Fabric, working closely with an established, globally distributed development team.
- Understand the application architecture, tools, and the processes, established and make recommendations to meet needs by application.
- Ensure that data is efficiently and accurately processed from source systems into the Microsoft Fabric data layer and transformed across the various levels of transformation inside Fabric.
- Develop transformation logic to clean, enrich, aggregate, and otherwise modify data to meet business needs. Implement data validation and error handling procedures to ensure data integrity.
- Implement processes to load processed data into Microsoft Fabric data storage or previously worked in technologies like Azure Synapse, Azure Data Lake, etc. Optimize data loading processes to ensure performance and scalability.
- Design and deploy ETL pipelines and notebooks using Python and Pyspark within Microsoft Fabric.
- Collaborate with data architects to design the data architecture and models that align with the application’s requirements. Implement data schemas, star/snowflake schema designs for data warehouses, and OLAP cubes, as necessary.
- Optimize ETL processes for performance and resource utilization. Implement indexing, partitioning, and other performance enhancements to handle large volumes of data efficiently within Fabric.
- Develop and execute unit tests, integration tests, and end-to-end tests for ETL processes in Fabric. Troubleshoot and resolve issues, making improvements and adjustments as needed.
- Validate data accuracy and consistency throughout the ETL lifecycle.
- Create and maintain comprehensive documentation for ETL processes, data sources, data flow diagrams, and data models.
- Ensure documentation is up-to-date and accessible to relevant stakeholders.
- Ensure all data processing complies with relevant policies and standards.
- Required Experience
- High proficiency in Microsoft Fabric and related ETL tools (e.g., Azure Data Factory)
- Knowledge of database systems (e.g., SQL Server, Azure SQL Database, Synapse Analytics) and understanding of data warehousing concepts and architecture.
- Experience with data modeling and schema design.
- Familiarity with programming languages used in ETL processes (e.g., Python, Pyspark).
- Strong understanding of data engineering principles, including data modeling, data transformation, and data optimization.
- Strong SQL skills for data extraction, transformation, and querying.
- Knowledge of accounting principles and logic is highly beneficial.
- Working knowledge of US accounting principles and accounting logic would be highly beneficial and desired.
- Desired Certifications
- Microsoft Certified: Fabric Analytics Engineer Associate (strongly preferred)
- Microsoft Certified: Azure Data Engineer Associate
- Microsoft Certified: Azure Developer Associate
- Microsoft Certified: Azure Solutions Architect Expert
- Responsibilities
- An experienced Microsoft Fabric ETL professional capable of integrating into KPMG’s existing processes and workflows.
- Collaborating with a global ETL and application development team to share knowledge, best practices, and ensure consistent development standards and to continually enhance development processes based on feedback and emerging industry trends.
- Utilizing modern development tools and frameworks currently in use by the KPMG development team (e.g., Azure Dev Ops).
- Ensuring timely reporting and proactive communication with the global development team leadership.
- Collaborate with business analysts and business stakeholders to understand data requirements, transformation rules, and relevant accounting business logic.
- Design, implement, and optimize ETL processes using Microsoft Fabric, working closely with an established, globally distributed development team.
- Understand the application architecture, tools, and the processes, established and make recommendations to meet needs by application.
- Ensure that data is efficiently and accurately processed from source systems into the Microsoft Fabric data layer and transformed across the various levels of transformation inside Fabric.
- Develop transformation logic to clean, enrich, aggregate, and otherwise modify data to meet business needs. Implement data validation and error handling procedures to ensure data integrity.
- Implement processes to load processed data into Microsoft Fabric data storage or previously worked in technologies like Azure Synapse, Azure Data Lake, etc. Optimize data loading processes to ensure performance and scalability.
- Design and deploy ETL pipelines and notebooks using Python and Pyspark within Microsoft Fabric.
- Collaborate with data architects to design the data architecture and models that align with the application’s requirements. Implement data schemas, star/snowflake schema designs for data warehouses, and OLAP cubes, as necessary.
- Optimize ETL processes for performance and resource utilization. Implement indexing, partitioning, and other performance enhancements to handle large volumes of data efficiently within Fabric.
- Develop and execute unit tests, integration tests, and end-to-end tests for ETL processes in Fabric. Troubleshoot and resolve issues, making improvements and adjustments as needed.
- Validate data accuracy and consistency throughout the ETL lifecycle.
- Create and maintain comprehensive documentation for ETL processes, data sources, data flow diagrams, and data models.
- Ensure documentation is up-to-date and accessible to relevant stakeholders.
- Ensure all data processing complies with relevant policies and standards.
- Required Experience
- High proficiency in Microsoft Fabric and related ETL tools (e.g., Azure Data Factory)
- Knowledge of database systems (e.g., SQL Server, Azure SQL Database, Synapse Analytics) and understanding of data warehousing concepts and architecture.
- Experience with data modeling and schema design.
- Familiarity with programming languages used in ETL processes (e.g., Python, Pyspark).
- Strong understanding of data engineering principles, including data modeling, data transformation, and data optimization.
- Strong SQL skills for data extraction, transformation, and querying.
- Knowledge of accounting principles and logic is highly beneficial.
- Working knowledge of US accounting principles and accounting logic would be highly beneficial and desired.
- Desired Certifications
- Microsoft Certified: Fabric Analytics Engineer Associate (strongly preferred)
- Microsoft Certified: Azure Data Engineer Associate
- Microsoft Certified: Azure Developer Associate
- Microsoft Certified: Azure Solutions Architect Expert
* Salary range is an estimate based on our AI, ML, Data Science Salary Index 💰
Job stats:
0
0
0
Categories:
Engineering Jobs
Leadership Jobs
Tags: Architecture Azure Data Warehousing Engineering ETL OLAP Pipelines PySpark Python Snowflake SQL
Region:
Asia/Pacific
Country:
India
More jobs like this
Explore more career opportunities
Find even more open roles below ordered by popularity of job title or skills/products/technologies used.
Data Scientist II jobsData Engineer II jobsSr. Data Engineer jobsBusiness Intelligence Analyst jobsPrincipal Data Engineer jobsStaff Data Scientist jobsStaff Machine Learning Engineer jobsData Science Manager jobsData Manager jobsPrincipal Software Engineer jobsData Science Intern jobsBusiness Data Analyst jobsJunior Data Analyst jobsData Analyst Intern jobsSoftware Engineer II jobsData Specialist jobsSr. Data Scientist jobsDevOps Engineer jobsLead Data Analyst jobsResearch Scientist jobsStaff Software Engineer jobsData Engineer III jobsAI/ML Engineer jobsSenior Backend Engineer jobsBI Analyst jobs
Git jobsAirflow jobsOpen Source jobsEconomics jobsLinux jobsKafka jobsComputer Vision jobsGoogle Cloud jobsJavaScript jobsMLOps jobsPhysics jobsNoSQL jobsTerraform jobsKPIs jobsData Warehousing jobsRDBMS jobsPostgreSQL jobsBanking jobsScikit-learn jobsHadoop jobsScala jobsGitHub jobsData warehouse jobsStreaming jobsPandas jobs
R&D jobsClassification jobsBigQuery jobsOracle jobsDistributed Systems jobsCX jobsPySpark jobsdbt jobsLooker jobsReact jobsScrum jobsRAG jobsMicroservices jobsJira jobsRobotics jobsSAS jobsRedshift jobsGPT jobsIndustrial jobsELT jobsNumPy jobsPrompt engineering jobsMySQL jobsData Mining jobsData strategy jobs