Data Engineer
India - Hyderabad
Amgen
Amgen is committed to unlocking the potential of biology for patients suffering from serious illnesses by discovering, developing, manufacturing and delivering innovative human therapeutics.Career Category
Information SystemsJob Description
Join Amgen’s Mission of Serving Patients
At Amgen, if you feel like you’re part of something bigger, it’s because you are. Our shared mission—to serve patients living with serious illnesses—drives all that we do.
Since 1980, we’ve helped pioneer the world of biotech in our fight against the world’s toughest diseases. With our focus on four therapeutic areas –Oncology, Inflammation, General Medicine, and Rare Disease– we reach millions of patients each year. As a member of the Amgen team, you’ll help make a lasting impact on the lives of patients as we research, manufacture, and deliver innovative medicines to help people live longer, fuller happier lives.
Our award-winning culture is collaborative, innovative, and science based. If you have a passion for challenges and the opportunities that lay within them, you’ll thrive as part of the Amgen team. Join us and transform the lives of patients while transforming your career.
Data Engineer
What you will do
Let’s do this. Let’s change the world. We are seeking an experienced Data Engineer to work on a GxP platform supporting the Regulatory Submission system and Clinical Trial Registry system. The role is responsible for designing, building, maintaining, analyzing, and interpreting data to deliver actionable insights that drive business decisions. This role involves working with large datasets, developing reports, supporting and implementing data governance initiatives and visualizing data to ensure data is accessible, reliable, and efficiently managed. The ideal candidate has strong technical skills, experience with big data technologies, and a deep understanding of data architecture and ETL processes
Roles & Responsibilities:
Design, develop, and maintain data solutions for data generation, collection, and processing
Be a key team member that assists in design and development of the data pipeline
Create data pipelines and ensure data quality by implementing ETL processes to migrate and deploy data across systems
Contribute to the design, development, and implementation of data pipelines, ETL/ELT processes, and data integration solutions
Take ownership of data pipeline projects from inception to deployment, manage scope, timelines, and risks
Collaborate with multi-functional teams to understand data requirements and design solutions that meet business needs
Develop and maintain data models, data dictionaries, and other documentation to ensure data accuracy and consistency
Implement data security and privacy measures to protect sensitive data
Leverage cloud platforms (AWS preferred) to build scalable and efficient data solutions
Collaborate and communicate effectively with product teams
Collaborate with Data Architects, Business SMEs, and Data Scientists to design and develop end-to-end data pipelines to meet fast-paced business needs across geographic regions
Identify and resolve complex data-related challenges
Adhere to standard methodologies for coding, testing, and designing reusable code/component
Explore new tools and technologies that will help to improve ETL platform performance
Participate in sprint planning meetings and provide estimations on technical implementation
Work with data engineers on data quality assessment, data cleansing and data analytics
Share and discuss findings with team members practicing SAFe Agile delivery model
Work as a Data Engineer for a team that uses Cloud and Big Data technologies to design, develop, implement and maintain solutions to support the R&D functional area.
Overall management of the Enterprise Data Lake on AWS environment to ensure that the service delivery is cost effective and business SLAs around uptime, performance and capacity are met.
Proactively work on challenging data integration problems by implementing optimal ETL patterns, frameworks for structured and unstructured data.
Automate and Optimize data pipeline and framework for easier and cost-effective development process.
Advice and support project teams (project managers, architects, business analysts, and developers) on cloud platforms (AWS, Databricks preferred), tools, technology, and methodology related to the design, build scalable, efficient and maintain Data Lake and other Big Data solutions
Experience developing in an Agile development environment, and comfortable with Agile terminology and ceremonies.
Stay up to date with the latest data technologies and trends.
What we expect of you
We are all different, yet we all use our unique contributions to serve patients.
Basic Qualifications:
Master’s degree and 1 to 3 years of Computer Science, IT or related field experience OR
Bachelor’s degree and 3 to 5 years of Computer Science, IT or related field experience OR
Diploma and 7 to 9 years of Computer Science, IT or related field experience
Have 3-5 years of experience in the Pharmaceutical Industry
Have 3-5 years of experience in Mulesoft development
Hands-on experience with big data technologies and platforms, such as Databricks, Apache Spark (PySpark, SparkSQL), workflow orchestration, performance tuning on big data processing
Hands-on experience with various Python/R packages for EDA, feature engineering, and machine learning model training
Proficiency in data analysis tools (eg. SQL) and experience with data visualization tools
Excellent problem-solving skills and the ability to work with large, complex datasets
Solid understanding of data governance frameworks, tools, and standard methodologies.
Knowledge of data protection and pharmaceutical regulations and compliance requirements (e.g., GxP, GDPR, CCPA)
Demonstrated hands-on experience with AWS cloud platform and its technologies like EC2, RDS, S3, Redshift, and IAM roles.
Extensive hands-on experience of working on Data Ingestion methods such as Batch, API and Streaming.
Knowledge of data protection regulations and compliance requirements (e.g., GDPR, CCPA).
Demonstrated experience of performing Data Integrations using Mulesoft.
Solid understanding of ETL, Data Modeling and Data Warehousing concepts.
Ability to work independently with little supervision
Ability to effectively present information to collaborators, and respond to questions to their questions
Preferred Qualifications:
Knowledge for clinical data in the pharmaceutical industry
Knowledge of CT.gov and EUCTR.gov portals
Knowledge of the Disclose application from Citeline
Experience with ETL tools such as Apache Spark, and various Python packages related to data processing, machine learning model development
Solid understanding of data modeling, data warehousing, and data integration concepts
Knowledge of Python/R, Databricks, SageMaker, cloud data platforms
Proficiency with Data Orchestration tools like Kubernetes, Docker etc.
Familiarity with SQL/NOSQL database, Vector Database for Large Language Models.
Professional Certifications:
SAFe® for Teams certification (preferred)
Databricks Certification(Preferred)
AWS Certified Data Engineer(Preferred)
Soft Skills:
Excellent critical-thinking and problem-solving skills
Strong communication and collaboration skills
Demonstrated awareness of how to function in a team setting
Demonstrated presentation skills
Shift Information:
This position requires you to work a later shift and may be assigned a second or third shift schedule. Candidates must be willing and able to work during evening or night shifts, as required based on business requirements.
What you can expect of us
As we work to develop treatments that take care of others, we also work to care for your professional and personal growth and well-being. From our competitive benefits to our collaborative culture, we’ll support your journey every step of the way.
In addition to the base salary, Amgen offers competitive and comprehensive Total Rewards Plans that are aligned with local industry standards.
Apply now and make a lasting impact with the Amgen team.
careers.amgen.com
As an organization dedicated to improving the quality of life for people around the world, Amgen fosters an inclusive environment of diverse, ethical, committed and highly accomplished people who respect each other and live the Amgen values to continue advancing science to serve patients. Together, we compete in the fight against serious disease.
Amgen is an Equal Opportunity employer and will consider all qualified applicants for employment without regard to race, color, religion, sex, sexual orientation, gender identity, national origin, protected veteran status, disability status, or any other basis protected by applicable law.
We will ensure that individuals with disabilities are provided reasonable accommodation to participate in the job application or interview process, to perform essential job functions, and to receive other benefits and privileges of employment. Please contact us to request accommodation.
.* Salary range is an estimate based on our AI, ML, Data Science Salary Index 💰
Tags: Agile APIs Architecture AWS Big Data Computer Science Data analysis Data Analytics Databricks Data governance Data pipelines Data quality Data visualization Data Warehousing Docker EC2 EDA ELT Engineering ETL Feature engineering Kubernetes LLMs Machine Learning ML models Model training NoSQL Pharma Pipelines Privacy PySpark Python R R&D Redshift Research SageMaker Security Spark SQL Streaming Testing Unstructured data
Perks/benefits: Career development Competitive pay
More jobs like this
Explore more career opportunities
Find even more open roles below ordered by popularity of job title or skills/products/technologies used.