Senior Data Engineer
Orlando, Florida, United States - Remote
Worth AI
Complete applications effortlessly with Worth Pre-Filling. Our AI-powered technology guarantees real-time accuracy and eliminates manual data entry.Worth AI, a leader in the computer software industry, is looking for a talented and experienced Senior Data Engineer to join their innovative team. At Worth AI, we are on a mission to revolutionize decision-making with the power of artificial intelligence while fostering an environment of collaboration, and adaptability, aiming to make a meaningful impact in the tech landscape.. Our team values include extreme ownership, one team and creating reaving fans both for our employees and customers.
As a Senior Data Engineer, you will lead the design and development of data services and platforms that power our AI-driven products. You'll focus on creating well-structured, validated APIs and event-driven pipelines, enabling scalable, secure, and maintainable data workflows. This is a backend-heavy role ideal for engineers who thrive on clean architecture, automation, and cross-functional delivery.
Responsibilities
- Design and build production-grade **FastAPI services** to serve, validate, and enrich data for ML and analytics use cases
- Create and maintain **asynchronous event-driven pipelines** usingĀ
- *Apache Kafka**, ensuring reliable and scalable communication across microservices
- Define and enforce structured data contracts using **Pydantic** and OpenAPI standards
- Develop robust, containerized data services with **Docker** and deploy them using modern cloud-native tooling
- Build and optimize analytical models and data flows in **Amazon Redshift** for business-critical reporting and data science consumption
- Collaborate with data scientists, ML engineers, and backend developers to streamline data sourcing, transformation, and model inference
- Own the lifecycle of data services ā including monitoring, observability, testing, and deployment pipelines
- Maintain rigorous standards around data privacy, schema governance, and system performance
- Design, build, code and maintain large-scale data processing systems and architectures that support AI initiatives.
- Develop and implement data pipelines and ETL processes to ingest, transform, and load data from various sources.
- Design and optimize databases and data storage solutions for high performance and scalability.
- Collaborate with cross-functional teams to understand data requirements and ensure data quality and integrity.
- Implement data governance and data security measures to protect sensitive data.
- Monitor and troubleshoot data infrastructure and pipeline issues in a timely manner.
- Stay up-to-date with the latest trends and technologies in data engineering and recommend improvements to enhance the company's data capabilities.
Requirements
- 7+ years of professional experience in backend-focused data engineering or platform development
- Strong proficiency in **Python**, with hands-on experience using **FastAPI**, **Pydantic**, and asynchronous programming patterns
- Deep understanding of **event-driven architectures** and experience with **Kafka** (producers, consumers, schema evolution, retries, etc.)
- Experience designing and deploying **containerized services** with **Docker** (Kubernetes or Fargate experience is a plus)
- Proficiency in SQL and experience with modern cloud data warehouses, preferably **Amazon Redshift**
- Familiarity with cloud services (preferably AWS), including CI/CD, infrastructure-as-code, and observability tooling
- Experience integrating third-party APIs and working with versioned schema contracts
- Strong communication and collaboration skills, especially in cross-functional and agile teams
- Experience working with ML engineers to operationalize models (e.g., batch scoring, online inference, data validation at model boundaries)
- In-depth knowledge of data modeling, data warehousing, and database design principles.
- Strong programming skills in Python, SQL, and other relevant languages.
- Experience with relational and NoSQL databases, such as PostgreSQL, MySQL, MongoDB
- Proficiency in data integration and ETL tools, such as Apache Kafka, Apache Airflow, or Informatica.
- Familiarity with big data processing frameworks, such as Hadoop, Spark, or Flink.
- Knowledge of cloud platforms, such as AWS, Azure, or GCP, and experience with data storage and processing services in the cloud.
- Understanding of data governance, data privacy, and data security best practices.
- Strong problem-solving and troubleshooting skills, with a focus on data quality and system performance.
- Excellent communication and collaboration skills to work effectively with cross-functional teams.
- Prior collaborative work with data scientists or machine learning professionals with respect to sourcing, processing and scaling both input and output data
- Comfortable going through documentation of third-party APIās and identifying best procedures for integrating data from APIās into broader ETL processesĀ
Benefits
- Health Care Plan (Medical, Dental & Vision)
- Retirement Plan (401k, IRA)
- Life Insurance
- Unlimited Paid Time Off
- 9 paid Holidays
- Family Leave
- Work From Home
- Free Food & Snacks (Access to Industrious Co-working Membership!)
- Wellness Resources
* Salary range is an estimate based on our AI, ML, Data Science Salary Index š°
Tags: Agile Airflow APIs Architecture AWS Azure Big Data CI/CD Data governance Data pipelines Data quality Data Warehousing Docker Engineering ETL FastAPI Flink GCP Hadoop Informatica Kafka Kubernetes Machine Learning Microservices Model inference MongoDB MySQL NoSQL Pipelines PostgreSQL Privacy Python Redshift Security Spark SQL Testing
Perks/benefits: 401(k) matching Career development Health care Medical leave Unlimited paid time off Wellness
More jobs like this
Explore more career opportunities
Find even more open roles below ordered by popularity of job title or skills/products/technologies used.