Data Engineer
Bengaluru, Karnataka, India
2070 Health
2070 Health is India’s first Venture Studio which is focusing on building patient-centric healthcare companies from scratch.2070 Health is India’s first venture studio dedicated to building patient-centric healthcare companies from scratch. By combining innovative ideas, exceptional founders, and a robust innovation platform, 2070 Health aims to transform the healthcare landscape in India. Through data-driven insights, we strive to improve patient outcomes and streamline healthcare delivery. Learn more at 2070health.com.
Job Summary:
As a Data Engineer at 2070 Health, you will design, build, and maintain robust data pipelines and infrastructure. You will work closely with cross-functional teams to enable data-driven decisions through efficient data storage, transformation, and visualization. The ideal candidate is proficient in Python, MySQL, AWS S3, Amazon Redshift, Excel, and Google Sheets, with experience in tools like dbt, Airbyte, and Apache Superset being a bonus.
Responsibilities:
Data Pipeline Development:
- Design, develop, and maintain scalable ETL (Extract, Transform, Load) processes.
- Build and maintain data transformation and automation workflows.
- Optimize data storage and retrieval processes using MySQL and Amazon Redshift.
- Ensure data accuracy, consistency, and reliability across various data sources.
Database Management:
- Manage and maintain MySQL and Redshift databases, including writing and optimizing SQL queries.
- Implement best practices for database performance, security, and scalability.
- Develop stored procedures, views, and functions to support business reporting needs.
Data Visualization and Reporting:
- Support data-driven decision-making by preparing reports and dashboards using Excel and Google Sheets.
- Create insightful visualizations using Apache Superset or other BI tools (bonus).
- Collaborate with stakeholders to understand reporting requirements and deliver timely insights.
Collaboration and Documentation:
- Work closely with Product Managers, Analysts, and other Engineers to understand data needs.
- Document data architecture, pipelines, and processes clearly and comprehensively.
- Support troubleshooting and resolution of data-related issues.
Continuous Improvement:
- Identify and implement improvements in data processes to enhance efficiency and reliability.
- Stay current with industry trends in data engineering and suggest relevant tools or technologies.
Requirements
- Bachelor’s degree in Computer Science, Engineering, Data Science, or a related field (or equivalent experience).
- 2-4 years of experience as a Data Engineer or in a similar role.
- Strong proficiency in Python, SQL and experience working with MySQL and Amazon Redshift.
- Experience building and maintaining ETL pipelines with tools like Airbyte.
- Proficiency in data transformation using dbt.
- Expertise in Excel and Google Sheets for data analysis and reporting.
- Solid understanding of database design, data modeling, and query optimization.
Preferred:
- Experience with data visualization tools like Apache Superset (bonus).
- Experience with open source tools like Dagster, DBT, Airbyte, Kafka (bonus).
- Familiarity with cloud platforms (AWS, GCP).
- Knowledge of scripting languages (e.g., Python) for automation.
- Excellent problem-solving, communication, and documentation skills.
- Experience working in agile environments.
* Salary range is an estimate based on our AI, ML, Data Science Salary Index 💰
Tags: Agile Architecture AWS Computer Science Dagster Data analysis Data pipelines Data visualization dbt Engineering ETL Excel GCP Kafka MySQL Open Source Pipelines Python Redshift Security SQL Superset
More jobs like this
Explore more career opportunities
Find even more open roles below ordered by popularity of job title or skills/products/technologies used.