Consultant Data Engineering - Connected Medicine
IN: Bengaluru - LCCI, India
Eli Lilly and Company
Lilly is a medicine company turning science into healing to make life better for people around the world.At Lilly, we unite caring with discovery to make life better for people around the world. We are a global healthcare leader headquartered in Indianapolis, Indiana. Our employees around the world work to discover and bring life-changing medicines to those who need them, improve the understanding and management of disease, and give back to our communities through philanthropy and volunteerism. We give our best effort to our work, and we put people first. We’re looking for people who are determined to make life better for people around the world.
For more than 140 years, Eli Lilly and Company has been dedicated to the discovery and development of medicines to address the health care needs of people around the world. Driven by the broader, consumer-focused digital revolution and in anticipation of this same revolution in health care, Lilly is now further expanding the definition of product to include the design, build, integration and orchestration of digitally enabled health care interactions and has established a Connected Medicine Business Unit to accomplish these goals in the multiple therapeutic areas.
More specifically, we are interested in using digital technologies to improve the health outcomes of those using our products and are formalizing our ecosystem of capabilities to orchestrate personalized engagement to Patients, Coupling our medicines, devices, and digital solutions with consumer health solutions and health systems, we believe we can partner with providers and payers to enable significant improvements in health outcomes for our customers.
The LCCI Connected Medicine is team comprising of creative problem solvers who support achievement of the brand’s goals, while always striving for process improvement and innovation.
The role will be responsible for setting up the data warehouses necessary to handle large volumes of data, create meaningful analyses, and deliver recommendations to leadership.
Core Responsibilities
- Create and maintain optimal data pipeline architecture ETL/ ELT into structured data
- Assemble large, complex data sets that meet business requirements and create and maintain multi-dimensional modelling like Star Schema and Snowflake Schema, normalization, de-normalization, joining of datasets.
- Expert level experience in creating a scalable data warehouse including Fact tables, Dimensional tables and ingest datasets into cloud based tools.
- Identify, design, and implement internal process improvements including automating manual processes, optimizing data delivery and re-designing infrastructure for greater scalability.
- Collaborate with stakeholders to ensure seamless integration of data with internal data marts, enhancing advanced reporting
- Setup and maintain data ingestion, streaming, scheduling, and job monitoring automation using AWS services. Setup Lambda, code pipeline (CI/CD), Glue, S3, Redshift, Power BI needs to be maintained for uninterrupted automation.
- Build analytics tools that utilize the data pipeline to provide actionable insight into customer acquisition, operational efficiency, and other key business performance metrics.
- Work with stakeholders to assist with data-related technical issues and support their data infrastructure needs.
- Utilize GitHub for version control, code collaboration, and repository management. Implement best practices for code reviews, branching strategies, and continuous integration.
- Create data tools for analytics and data scientist team members that assist them in building and optimizing our product into an innovative industry leader
- Ensure data privacy and compliance with relevant regulations (e.g., GDPR) when handling customer data.
- Maintain data quality and consistency within the application, addressing data-related issues as they arise.
Required
- 7-10 years of relevant experience
- Advanced working SQL knowledge and experience working with relational databases, query authoring (SQL) as well as working familiarity with a variety of databases and Cloud Data warehouse like AWS Redshift
- Experience in creating scalable, efficient schema designs to support diverse business needs.
- Experience with database normalization, schema evolution, and maintaining data integrity
- Proactively share best practices, contributing to team knowledge and improving schema design transitions.
- Develop data models, create dimensions and facts, and establish views and procedures to enable automation programmability.
- Collaborate effectively with cross-functional teams to gather requirements, incorporate feedback, and align analytical work with business objectives
- Prior Data Modelling, OLAP cube modelling
- Data compression into PARQUET to improve processing and finetuning SQL programming skills.
- Experience building and optimizing “big data” data pipelines, architectures and data sets.
- Experience performing root cause analysis on internal and external data and processes to answer specific business questions and identify opportunities for improvement.
- Experience with manipulating, processing and extracting value from large disconnected unrelated datasets
- Strong analytic skills related to working with structured and unstructured datasets.
- Working knowledge of message queuing, stream processing, and highly scalable “big data” stores.
- Experience supporting and working with cross-functional teams and Global IT.
- Familiarity of working in an agile based working models.
Preferred Qualifications/Expertise
- Experience with relational SQL and NoSQL databases, especially AWS Redshift.
- Experience with AWS cloud services Preferable: S3, EC2, Lambda, Glue, EMR, Code pipeline highly preferred. Experience with similar services on another platform would also be considered.
Education:
- Bachelor’s or master’s degree on Technology and Computer Science background
Lilly is dedicated to helping individuals with disabilities to actively engage in the workforce, ensuring equal opportunities when vying for positions. If you require accommodation to submit a resume for a position at Lilly, please complete the accommodation request form (https://careers.lilly.com/us/en/workplace-accommodation) for further assistance. Please note this is for individuals to request an accommodation as part of the application process and any other correspondence will not receive a response.
Lilly does not discriminate on the basis of age, race, color, religion, gender, sexual orientation, gender identity, gender expression, national origin, protected veteran status, disability or any other legally protected status.
#WeAreLilly
* Salary range is an estimate based on our AI, ML, Data Science Salary Index 💰
Tags: Agile Architecture AWS Big Data CI/CD Computer Science Data pipelines Data quality Data warehouse EC2 ELT Engineering ETL GitHub Lambda NoSQL OLAP Parquet Pipelines Power BI Privacy RDBMS Redshift Snowflake SQL Streaming
More jobs like this
Explore more career opportunities
Find even more open roles below ordered by popularity of job title or skills/products/technologies used.