Data Architect
Pune DIA, India
Roche
As a pioneer in healthcare, we have been committed to improving lives since the company was founded in 1896 in Basel, Switzerland. Today, Roche creates innovative medicines and diagnostic tests that help millions of patients globally.At Roche you can show up as yourself, embraced for the unique qualities you bring. Our culture encourages personal expression, open dialogue, and genuine connections, where you are valued, accepted and respected for who you are, allowing you to thrive both personally and professionally. This is how we aim to prevent, stop and cure diseases and ensure everyone has access to healthcare today and for generations to come. Join Roche, where every voice matters.
The Position
In Roche Informatics, we build on Roche’s 125-year history as one of the world’s largest biotech companies, globally recognized for providing transformative innovative solutions across major disease areas. We combine human capabilities with cutting-edge technological innovations to do now what our patients need next. Our commitment to our patients’ needs motivates us to deliver technology that evolves the practice of medicine.
Be part of our inclusive team at Roche Informatics, where we’re driven by a shared passion for technological novelties and optimal IT solutions.
Position Overview
We are seeking an experienced ETL Architect to design, develop, and optimize data extraction, transformation, and loading (ETL) solutions and to work closely with multi-disciplinary and multi-cultural teams to build structured, high-quality data solutions. The person may be leading technical squads. These solutions will be leveraged across Enterprise , Pharma and Diagnostics solutions to help our teams fulfill our mission: to do now what patients need next.
This role requires deep expertise in Python, AWS Cloud, and ETL tools to build and maintain scalable data pipelines and architectures. The ETL Architect will work closely with cross-functional teams to ensure efficient data integration, storage, and accessibility for business intelligence and analytics.
Key Responsibilities
ETL Design & Development: Architect and implement high-performance ETL pipelines using AWS cloud services, Snowflake, and ETL tools such as Talend, Dbt, Informatica, ADF etc.
Data Architecture: Design and implement scalable, efficient, and cloud-native data architectures.
Data Integration & Flow: Ensure seamless data integration across multiple source systems, leveraging AWS Glue, Snowflake, and other ETL tools.
Performance Optimization: Monitor and tune ETL processes for performance, scalability, and cost-effectiveness.
Governance & Security: Establish and enforce data quality, governance, and security standards for ETL processes.
Collaboration: Work with data engineers, analysts, and business stakeholders to define data requirements and ensure effective solutions.
Documentation & Best Practices: Maintain comprehensive documentation and promote best practices for ETL development and data transformation.
Troubleshooting & Support: Diagnose and resolve performance issues, failures, and bottlenecks in ETL processes.
Required Qualifications
Education: Bachelor's or Master’s degree in Computer Science, Information Technology, Data Engineering, or related field.
Experience:
6+ years of experience in ETL development, with 3+ years in an ETL architecture role.
Expertise in Snowflake or any MPP data warehouse (including Snowflake data modeling, optimization, and security best practices).
Strong experience with AWS Cloud services, especially AWS Glue, AWS Lambda, S3, Redshift, and IAM or Azure/GCP cloud services.
Proficiency in ETL tools such as Informatica, Talend, Apache NiFi, SSIS, or DataStage.
Strong SQL skills and experience with relational and NoSQL databases.
Experience in API integrations
Proficiency in scripting languages (Python, Shell, PowerShell) for automation.
Prior experience in Pharmaceutical or Diagnostics or healthcare domain is a plus.
Soft Skills
Strong analytical and problem-solving abilities.
Excellent communication and documentation skills.
Ability to work collaboratively in a fast-paced, cloud-first environment.
Preferred Qualifications
Certifications in AWS, Snowflake, or ETL tools.
Experience in real-time data streaming, microservices-based architectures, and DevOps for data pipelines.
Knowledge of data governance, compliance (GDPR, HIPAA), and security best practices.
A healthier future drives us to innovate. Together, more than 100’000 employees across the globe are dedicated to advance science, ensuring everyone has access to healthcare today and for generations to come. Our efforts result in more than 26 million people treated with our medicines and over 30 billion tests conducted using our Diagnostics products. We empower each other to explore new possibilities, foster creativity, and keep our ambitions high, so we can deliver life-changing healthcare solutions that make a global impact.
Let’s build a healthier future, together.
Roche is an Equal Opportunity Employer.
* Salary range is an estimate based on our AI, ML, Data Science Salary Index 💰
Tags: APIs Architecture AWS AWS Glue Azure Business Intelligence Computer Science Data governance Data pipelines Data quality Data warehouse dbt DevOps Engineering ETL GCP Informatica Lambda Microservices MPP NiFi NoSQL Pharma Pipelines Python Redshift Security Snowflake SQL SSIS Streaming Talend
More jobs like this
Explore more career opportunities
Find even more open roles below ordered by popularity of job title or skills/products/technologies used.