Data Engineer

Illinois

CDC Foundation

The CDC Foundation is a global nonprofit, managing public health programs that impact chronic and infectious diseases and emergency threats like COVID-19.

View all jobs at CDC Foundation

Apply now Apply later

The CDC Foundation helps the Centers for Disease Control and Prevention (CDC) save and improve lives by unleashing the power of collaboration between CDC, philanthropies, corporations, organizations and individuals to protect the health, safety and security of America and the world. The CDC Foundation is the go-to nonprofit authorized by Congress to mobilize philanthropic partners and private-sector resources to support CDC’s critical health protection mission. Since 1995, the CDC Foundation has raised over $1.9 billion and launched more than 1,300 programs impacting a variety of health threats from chronic disease conditions including cardiovascular disease and cancer, to infectious diseases like rotavirus and HIV, to emergency responses, including COVID-19 and Ebola. The CDC Foundation managed hundreds of programs in the United States and in more than 90 countries last year. Visit www.cdcfoundation.org for more information.  
Job HighlightsLocation: Remote, must be based in the United StatesSalary Range: $103,500-$143,500 per year, plus benefits. Individual salary offers will be based on experience and qualifications unique to each candidate. Position Type: Grant funded, limited-term opportunityPosition End Date: June 30, 2025 Hours of Work:  Individual must routinely be available during core department business hours to collaborate with Illinois Department of Public Health staff.
OverviewThe Data Engineer will play a crucial role in advancing the CDC Foundation's mission by designing, building, and maintaining data infrastructure for a public health organization. This role is aligned to the Workforce Acceleration Initiative (WAI). WAI is a federally funded CDC Foundation program with the goal of helping the nation’s public health agencies by providing them with the technology and data experts they need to accelerate their information system improvements. Working within the Illinois Department of Public Health, to build, optimize, and manage cloud-based data pipelines and ETL processes on the Snowflake platform. This role will focus on implementing high-performance data systems, ensuring the reliability and scalability of our data infrastructure, and enabling robust analytics and insights for the organization. The ideal candidate will have hands-on experience with Snowflake, cloud environments, and data engineering best practices. The Data Engineer will be hired by the CDC Foundation and assigned to the Illinois Department of Public Health. This position is eligible for a fully remote work arrangement for U.S. based candidates.

Responsibilities

  • Develop a detailed plan for database migration, ETL processes, and data processing applications.
  • Design, build, and manage ETL/ELT processes and data pipelines on the Snowflake platform, ensuring the movement of large datasets between various data sources. Develop efficient, scalable data architectures and implement Snowflake best practices, including partitioning, clustering, and query optimization for performance and cost. Collaborate with data scientists, analysts, and Local health departments to integrate diverse data sources into Snowflake, ensuring data is available for analytics and reporting.
  • Monitor data pipelines and systems for performance issues, costs, errors and anomalies, and implement solutions to address them.
  • Collaborate with the IT Security Team to conduct security and access testing. Implement security measures to protect sensitive information.
  • Collaborate with cross-functional teams to understand data requirements and design scalable solutions that meet business needs. Collaborate with Systems Architect on overall system health, focusing on data aspects and data warehouse. Collaborate with Systems Architect on infrastructure assessment, focusing on data aspects.
  • Implement and maintain ETL processes to ensure the accuracy, completeness, and consistency of data.
  • Design and manage data storage systems, including relational databases, NoSQL databases, and data warehouses.
  • Knowledgeable about industry trends, best practices, and emerging technologies in data engineering, and incorporating the trends into the organization's data infrastructure.
  • Provide technical guidance to other staff. Create and maintain clear documentation for ETL processes, data pipelines, data models, and infrastructure setups. Develop training materials and conduct online sessions on accessing and utilizing shared data.
  • Communicate effectively with partners at all levels of the organization to gather requirements, provide updates, and present findings.
  • Create a data governance framework for secure and compliant data sharing.
  • Establish successful connection migration plan for ETL processes and APIs between migrated applications and databases.
  • Implement automated processes for data extraction from source systems and loading into the data warehouse.
  • Migrate ETL processes and APIs to the cloud environment.

Qualifications

  • Bachelor's degree in Computer Science, Information Technology, Data Science, or a related field.
  • Minimum 5 years of relevant professional experience.
  • Proficiency in programming languages commonly used in data engineering, such as Python, Java, Scala, or SQL. Candidate should be able to implement data automations within existing frameworks as opposed to writing one off scripts.
  • Experience with big data technologies and frameworks like Hadoop, Spark, Kafka, and Flink.
  • High level of proficiency in Snowflake, including advanced features like Time Travel, Zero-Copy Cloning, and data sharing is required. Experience regarding engineering best practices such as source control, automated testing, continuous integration and deployment, and peer review.
  • Knowledge of data warehousing concepts and tools.Familiarity with data lake and lakehouse architectures.
  • Expertise in data modeling, ETL (Extract, Transform, Load) processes, and data integration techniques.
  • Familiarity with agile development methodologies, software design patterns, and best practices.
  • Strong analytical thinking and problem-solving abilities.
  • Excellent verbal and written communication skills, including the ability to convey technical concepts to non-technical partners effectively.
  • Flexibility to adapt to evolving project requirements and priorities.
  • Outstanding interpersonal and teamwork skills; and the ability to develop productive working relationships with colleagues and partners.
  • Experience working in a virtual environment with remote partners and teams
  • Proficiency in Microsoft Office.
Special NotesThis role is involved in a dynamic public health program. As such, roles and responsibilities are subject to change as situations evolve. Roles and responsibilities listed above may be expanded upon or updated to match priorities and needs, once written approval is received by the CDC Foundation in order to best support the public health programming.
All qualified applicants will receive consideration for employment and will not be discriminated against on the basis of race, color, religion, sex, national origin, age, mental or physical disabilities, veteran status, and all other characteristics protected by law.
We comply with all applicable laws including E.O. 11246 and the Vietnam Era Readjustment Assistance Act of 1974 governing employment practices and do not discriminate on the basis of any unlawful criteria in accordance with 41 C.F.R. §§ 60-300.5(a)(12) and 60-741.5(a)(7). As a federal government contractor, we take affirmative action on behalf of protected veterans.
The CDC Foundation is a smoke-free environment.  Relocation expenses are not included.
Apply now Apply later
  • Share this job via
  • 𝕏
  • or
Job stats:  0  0  0
Category: Engineering Jobs

Tags: Agile APIs Architecture Big Data Clustering Computer Science Data governance Data pipelines Data warehouse Data Warehousing ELT Engineering ETL Flink Hadoop Java Kafka Nonprofit NoSQL Pipelines Python R RDBMS Scala Security Snowflake Spark SQL Testing

Perks/benefits: Health care Relocation support

Regions: Remote/Anywhere North America
Country: United States

More jobs like this