Senior Data Engineer
Remote, US
Care Access
Care Access is a health research company that improves access to research generating new cures & treatments for diseasesWho We AreWe care. Our people are the engines behind our mission: to revolutionize access to clinical trials for the benefit of patients everywhere. We care for one another, find new ideas to accelerate medicine, and seed a long-term impact for generations.
Position OverviewWe are seeking an experienced and detail-oriented professional to join our team as a Sr. Data Engineer. In this pivotal role, you will be responsible for designing, developing, and maintaining robust data pipelines that ensure the reliable ingestion, transformation, and delivery of complex data (demographics, medical, financial, marketing, etc.) across systems. The ideal candidate will bring deep expertise in Databricks, SQL, and modern data engineering practices, along with strong collaboration skills to help drive excellence across our data infrastructure.
What You'll Be Working On (Duties include but not limited to):
- Data Engineering Strategy and Architecture:
- Design and implement scalable, reliable, and efficient data pipelines to support clinical, operational, and business needs.
- Develop and maintain architecture standards, reusable frameworks, and best practices across data engineering workflows.
- Build automated systems for data ingestion, transformation, and orchestration leveraging cloud-native and open-source tools.
- Data Infrastructure and Performance Optimization:
- Optimize data storage and processing in data lakes and cloud data warehouses (Azure, Databricks).
- Develop and monitor batch and streaming data processes to ensure data accuracy, consistency, and timeliness.
- Maintain documentation and lineage tracking across datasets and pipelines to support transparency and governance.
- Collaboration and Stakeholder Engagement:
- Work cross-functionally with analysts, data scientists, software engineers, and business stakeholders to understand data requirements and deliver fit-for-purpose data solutions.
- Review and refine work completed by other team members, ensuring quality and performance standards are met.
- Provide technical mentorship to junior team members and collaborate with contractors and third-party vendors to extend engineering capacity.
- Technology and Tools:
- Use Databricks, DBT, Azure Data Factory, and SQL to architect and deploy robust data engineering solutions.
- Integrate APIs, structured/unstructured data sources, and third-party systems into centralized data platforms.
- Evaluate and implement new technologies to enhance the scalability, observability, and automation of data operations.
- Other Responsibilities
- Continuous Improvement: Proactively suggest improvements to infrastructure, processes, and automation to improve system efficiency, reduce costs, and enhance performance.
Scope of Role:
- Autonomy of Role: Work is performed under limited supervision
- Direct Reports: No
Physical Requirements:
- This role requires 100% of work to be performed in a remote office environment and requires the ability to use keyboards and other computer equipment.
Travel Requirements:
- This is a remote position with less than 10% travel requirements. Occasional planned travel may be required as part of the role.
What You Bring (Knowledge, Skills, and Abilities):
- Strong expertise in Databricks, SQL, dbt, Python, and cloud data ecosystems such as Azure.
- Experience working with structured and semi-structured data from diverse domains.
- Familiarity with CI/CD pipelines, orchestration tools (e.g., Airflow, Azure Data Factory), and modern software engineering practices.
- Strong analytical and problem-solving skills, with the ability to address complex data challenges and drive toward scalable solutions.
Certifications, Education, and Experience:
- Bachelor’s or master’s degree in computer science, Information Systems, Engineering, or a related field.
- 5+ years of experience in data engineering with a proven track record of building cloud-based, production-grade data pipelines.
Benefits (US Full-Time Employees Only):
- Paid Time Off (PTO) and Company Paid Holidays
- 100% Employer paid medical, dental, and vision insurance plan options
- Health Savings Account and Flexible Spending Accounts
- Bi-weekly HSA employer contribution
- Company paid Short-Term Disability and Long-Term Disability
- 401(k) Retirement Plan, with Company Match
Employment StatementCare Access complies with all employment laws and regulations with respect to its employment practices, terms and conditions of employment, and pay equity and wages. Care Access does not engage in any unfair or forced labor practice and does not tolerate, under any circumstances, the use of any form of forced or involuntary labor, child labor, or human trafficking. This extends to suppliers, partners, or other third parties with whom Care Access does business. Care Access values and promotes the protection of human rights everywhere.
* Salary range is an estimate based on our AI, ML, Data Science Salary Index 💰
Tags: Airflow APIs Architecture Azure CI/CD Computer Science Databricks DataOps Data pipelines dbt Engineering Open Source Pipelines Python Research SQL Streaming Unstructured data
Perks/benefits: 401(k) matching Flex hours Flexible spending account Flex vacation Health care Insurance Transparency
More jobs like this
Explore more career opportunities
Find even more open roles below ordered by popularity of job title or skills/products/technologies used.