Assistant Manager- Data Engineer
Bangalore, Karnataka, India
- Responsibilities
- An experienced Microsoft Fabric ETL professional capable of integrating into KPMG’s existing processes and workflows.
- Collaborating with a global ETL and application development team to share knowledge, best practices, and ensure consistent development standards and to continually enhance development processes based on feedback and emerging industry trends.
- Utilizing modern development tools and frameworks currently in use by the KPMG development team (e.g., Azure Dev Ops).
- Ensuring timely reporting and proactive communication with the global development team leadership.
- Collaborate with business analysts and business stakeholders to understand data requirements, transformation rules, and relevant accounting business logic.
- Design, implement, and optimize ETL processes using Microsoft Fabric, working closely with an established, globally distributed development team.
- Understand the application architecture, tools, and the processes, established and make recommendations to meet needs by application.
- Ensure that data is efficiently and accurately processed from source systems into the Microsoft Fabric data layer and transformed across the various levels of transformation inside Fabric.
- Develop transformation logic to clean, enrich, aggregate, and otherwise modify data to meet business needs. Implement data validation and error handling procedures to ensure data integrity.
- Implement processes to load processed data into Microsoft Fabric data storage or previously worked in technologies like Azure Synapse, Azure Data Lake, etc. Optimize data loading processes to ensure performance and scalability.
- Design and deploy ETL pipelines and notebooks using Python and Pyspark within Microsoft Fabric.
- Collaborate with data architects to design the data architecture and models that align with the application’s requirements. Implement data schemas, star/snowflake schema designs for data warehouses, and OLAP cubes, as necessary.
- Optimize ETL processes for performance and resource utilization. Implement indexing, partitioning, and other performance enhancements to handle large volumes of data efficiently within Fabric.
- Develop and execute unit tests, integration tests, and end-to-end tests for ETL processes in Fabric. Troubleshoot and resolve issues, making improvements and adjustments as needed.
- Validate data accuracy and consistency throughout the ETL lifecycle.
- Create and maintain comprehensive documentation for ETL processes, data sources, data flow diagrams, and data models.
- Ensure documentation is up-to-date and accessible to relevant stakeholders.
- Ensure all data processing complies with relevant policies and standards.
- Required Experience
- High proficiency in Microsoft Fabric and related ETL tools (e.g., Azure Data Factory)
- Knowledge of database systems (e.g., SQL Server, Azure SQL Database, Synapse Analytics) and understanding of data warehousing concepts and architecture.
- Experience with data modeling and schema design.
- Familiarity with programming languages used in ETL processes (e.g., Python, Pyspark).
- Strong understanding of data engineering principles, including data modeling, data transformation, and data optimization.
- Strong SQL skills for data extraction, transformation, and querying.
- Knowledge of accounting principles and logic is highly beneficial.
- Working knowledge of US accounting principles and accounting logic would be highly beneficial and desired.
- Desired Certifications
- Microsoft Certified: Fabric Analytics Engineer Associate (strongly preferred)
- Microsoft Certified: Azure Data Engineer Associate
- Microsoft Certified: Azure Developer Associate
- Microsoft Certified: Azure Solutions Architect Expert
- Responsibilities
- An experienced Microsoft Fabric ETL professional capable of integrating into KPMG’s existing processes and workflows.
- Collaborating with a global ETL and application development team to share knowledge, best practices, and ensure consistent development standards and to continually enhance development processes based on feedback and emerging industry trends.
- Utilizing modern development tools and frameworks currently in use by the KPMG development team (e.g., Azure Dev Ops).
- Ensuring timely reporting and proactive communication with the global development team leadership.
- Collaborate with business analysts and business stakeholders to understand data requirements, transformation rules, and relevant accounting business logic.
- Design, implement, and optimize ETL processes using Microsoft Fabric, working closely with an established, globally distributed development team.
- Understand the application architecture, tools, and the processes, established and make recommendations to meet needs by application.
- Ensure that data is efficiently and accurately processed from source systems into the Microsoft Fabric data layer and transformed across the various levels of transformation inside Fabric.
- Develop transformation logic to clean, enrich, aggregate, and otherwise modify data to meet business needs. Implement data validation and error handling procedures to ensure data integrity.
- Implement processes to load processed data into Microsoft Fabric data storage or previously worked in technologies like Azure Synapse, Azure Data Lake, etc. Optimize data loading processes to ensure performance and scalability.
- Design and deploy ETL pipelines and notebooks using Python and Pyspark within Microsoft Fabric.
- Collaborate with data architects to design the data architecture and models that align with the application’s requirements. Implement data schemas, star/snowflake schema designs for data warehouses, and OLAP cubes, as necessary.
- Optimize ETL processes for performance and resource utilization. Implement indexing, partitioning, and other performance enhancements to handle large volumes of data efficiently within Fabric.
- Develop and execute unit tests, integration tests, and end-to-end tests for ETL processes in Fabric. Troubleshoot and resolve issues, making improvements and adjustments as needed.
- Validate data accuracy and consistency throughout the ETL lifecycle.
- Create and maintain comprehensive documentation for ETL processes, data sources, data flow diagrams, and data models.
- Ensure documentation is up-to-date and accessible to relevant stakeholders.
- Ensure all data processing complies with relevant policies and standards.
- Required Experience
- High proficiency in Microsoft Fabric and related ETL tools (e.g., Azure Data Factory)
- Knowledge of database systems (e.g., SQL Server, Azure SQL Database, Synapse Analytics) and understanding of data warehousing concepts and architecture.
- Experience with data modeling and schema design.
- Familiarity with programming languages used in ETL processes (e.g., Python, Pyspark).
- Strong understanding of data engineering principles, including data modeling, data transformation, and data optimization.
- Strong SQL skills for data extraction, transformation, and querying.
- Knowledge of accounting principles and logic is highly beneficial.
- Working knowledge of US accounting principles and accounting logic would be highly beneficial and desired.
- Desired Certifications
- Microsoft Certified: Fabric Analytics Engineer Associate (strongly preferred)
- Microsoft Certified: Azure Data Engineer Associate
- Microsoft Certified: Azure Developer Associate
- Microsoft Certified: Azure Solutions Architect Expert
* Salary range is an estimate based on our AI, ML, Data Science Salary Index 💰
Job stats:
0
0
0
Categories:
Engineering Jobs
Leadership Jobs
Tags: Architecture Azure Data Warehousing Engineering ETL OLAP Pipelines PySpark Python Snowflake SQL
Region:
Asia/Pacific
Country:
India
More jobs like this
Explore more career opportunities
Find even more open roles below ordered by popularity of job title or skills/products/technologies used.
Sr. Data Engineer jobsBusiness Intelligence Developer jobsPower BI Developer jobsBI Developer jobsStaff Data Scientist jobsStaff Machine Learning Engineer jobsPrincipal Software Engineer jobsData Science Intern jobsDevOps Engineer jobsJunior Data Analyst jobsData Science Manager jobsSoftware Engineer II jobsData Manager jobsData Analyst Intern jobsLead Data Analyst jobsStaff Software Engineer jobsBusiness Data Analyst jobsAI/ML Engineer jobsAccount Executive jobsSr. Data Scientist jobsData Specialist jobsData Governance Analyst jobsSenior Backend Engineer jobsBusiness Intelligence Analyst jobsData Engineer III jobs
Consulting jobsMLOps jobsAirflow jobsOpen Source jobsEconomics jobsLinux jobsKPIs jobsKafka jobsTerraform jobsJavaScript jobsGitHub jobsData Warehousing jobsPostgreSQL jobsRDBMS jobsNoSQL jobsScikit-learn jobsStreaming jobsComputer Vision jobsClassification jobsBanking jobsPrompt engineering jobsPhysics jobsGoogle Cloud jobsRAG jobsOracle jobs
Pandas jobsHadoop jobsdbt jobsBigQuery jobsScala jobsR&D jobsLooker jobsData warehouse jobsGPT jobsReact jobsScrum jobsLangChain jobsPySpark jobsDistributed Systems jobsELT jobsMicroservices jobsIndustrial jobsCX jobsJira jobsSAS jobsRedshift jobsOpenAI jobsModel training jobsTypeScript jobsJenkins jobs