Clinical Data Engineer

IN: Bengaluru - LCCI

Eli Lilly and Company

Lilly is a medicine company turning science into healing to make life better for people around the world.

View all jobs at Eli Lilly and Company

Apply now Apply later

At Lilly, we unite caring with discovery to make life better for people around the world. We are a global healthcare leader headquartered in Indianapolis, Indiana. Our employees around the world work to discover and bring life-changing medicines to those who need them, improve the understanding and management of disease, and give back to our communities through philanthropy and volunteerism. We give our best effort to our work, and we put people first. We’re looking for people who are determined to make life better for people around the world.

Job Summary

At Lilly, we unite caring with discovery to make life better for people around the world. We are a global healthcare leader headquartered in Indianapolis, Indiana. Our 39,000 employees work to discover and bring life-changing medicines to those who need them, improve the understanding and management of disease, and give back to our communities through philanthropy and volunteerism. We give our best effort to our work, and we put people first. We’re looking for people who are determined to make life better for people around the globe.

Medicines Development serves a strategic function within Lilly Research Laboratories (LRL). Development of most medicines requires about a billion dollars and a decade or more to bring to market. At Lilly, we’re investing to make it faster, better and cheaper, to the benefit of patients... and the Information and Digital Solutions organization is a critical partner in reaching success. Broad trends in science and healthcare are driving integration across Medicines Development, Research, Manufacturing, and global Business Units. New digital capabilities in the broader business model, both internally and externally, will require agile and creative evolution of the processes and digital solutions that support clinical and product development.  

This position is responsible for designing, developing and supporting AWS based data pipeline and modelling capabilities in support of LRL business needs. Ensure successful transition of applications to service management team through planning and knowledge transfer Responsible for system testing, ensuring effective resolution of defects, timely discussion around business issues and appropriate management of resources relevant to data and integration.

As a Data engineer, you are responsible for ensuring the availability and quality of data needed for analysis and business transactions. This includes data integration, acquisition, cleansing, harmonization and transforming raw data into curated datasets for data science, data discovery, and BI/analytics. Responsible for developing, constructing, testing and maintaining data sets and scalable data processing systems. 

Data engineers work closest with Data Architects and Data Scientists.  You will also work with business and IT groups beyond the data sphere, understanding the enterprise infrastructure and the many source systems.

Key capabilities in this role family include:

• Data Acquisition -  is the process of gathering and storing data in a location and format that it can be consumed for data preparation and/or downstream business uses.

• Data Preparation - is an iterative process for exploring, integrating, cleaning, validating and transforming raw data into curated datasets

• Data Publishing -  is the act of releasing data in consumable form for (re)use by others.

Note: All data engineer roles should have a foundational set of knowledge in:  communication, leadership, teamwork, problem solving skills, solution / blueprint definition, business acumen, architectural processes (e.g. blueprinting, reference architecture, governance, etc.), technical standards, project delivery, and industry knowledge.

Business Analysis and Technical Leadership 

  • Engages with business and proactively seeks opportunities to deliver business value.  
  • Understands business requirements and effectively translates business needs and process into technical terms, and vice versa.
  • Elicits and defines requirements 
  • Ensures appropriate business roles are engaged in solution execution. 
  • Participates in design reviews to ensure traceability of requirements. 
  • Networks with appropriate IT colleagues to determine solutions to meet business partners’ needs. 
  • Seeks opportunities to reuse existing processes and services to streamline support and implementation of key systems.  
  • Stay abreast of tools and technologies to influence IT strategy so that it provides best usage opportunities for business.
  • Ability to adapt quickly in a constantly changing environment.

Must have technical skills:

  • Bachelor’s degree in computer science, Information Technology, Management Information Systems or equivalent work experience. 
  • 7+ years of intensive development experience in the core tools and technologies like SQL, Python, AWS ( Lamda, Glue, S3, Redshift, Athena, IAM Roles & Policies) , PySpark used by the solution services team.
  • Architect and build high-performance and scalable data pipelines adhering to data lakehouse, data warehouse & data marts standards for optimal storage, retrieval and processing of data.
  • 3+ years of experience in Agile Development and code deployment using Github & CI-CD pipelines.
  • 2+ years of experience in job orchestration using Airflow.
  • Handson experience with Programming Languages and Big Data Technologies like, Python, Hadoop, Spark, Kafka.
  • Expertise in Database Management using Database Management: Relational Databases: MySQL, PostgreSQL, Oracle and NoSQL Databases: MongoDB, Cassandra, Redis.
  • Expertise in the design, data modelling, creation and management of large datasets/data models.
  • Must have working experience with AWS Cloud Platforms and good expertise in Event-driven architectures - AWS SNS/SQS, Apache Kafka, RabbitMQ and AWS ETL services (AWS Glue, AWS Data Pipeline, Redshift) Or Equivalent Azure Services.
  • Good experience on Data Pipeline Orchestration and Data Warehousing Solutions.
  • Experience/Literacy on latest AI/ML, LLM technologies (Open AI, Bedrock, Claude).
  • Ability to work with business owners to define key business requirements and convert to technical specifications.
  • Experience with security models and development on large data sets.
  • Ensure successful transition of applications to service management team through planning and knowledge transfer.
  • Responsible for system testing, ensuring effective resolution of defects, timely discussion around business issues and appropriate management of resources relevant to data and integration.
  • Partner with and influence on solution development to ensure understanding of data and technical direction for solutions as well as delivery.

Lilly is dedicated to helping individuals with disabilities to actively engage in the workforce, ensuring equal opportunities when vying for positions. If you require accommodation to submit a resume for a position at Lilly, please complete the accommodation request form (https://careers.lilly.com/us/en/workplace-accommodation) for further assistance. Please note this is for individuals to request an accommodation as part of the application process and any other correspondence will not receive a response.

Lilly does not discriminate on the basis of age, race, color, religion, gender, sexual orientation, gender identity, gender expression, national origin, protected veteran status, disability or any other legally protected status.

#WeAreLilly

Apply now Apply later

* Salary range is an estimate based on our AI, ML, Data Science Salary Index 💰

Job stats:  0  0  0
Category: Engineering Jobs

Tags: Agile Airflow Architecture Athena AWS AWS Glue Azure Big Data Cassandra Claude Computer Science Data pipelines Data warehouse Data Warehousing ETL GitHub Hadoop Kafka LLMs Machine Learning MongoDB MySQL NoSQL Oracle Pipelines PostgreSQL PySpark Python RabbitMQ RDBMS Redshift Research Security Spark SQL Testing

Region: Asia/Pacific
Country: India

More jobs like this