Senior Data Analytics Engineer
United States - Remote
Applications have closed
Position Overview:
As a Senior ETL Operations and Data Analytics Engineer, you will play a crucial role in our data-driven decision-making process. You will be responsible for designing, implementing, and maintaining ETL processes, ensuring data accuracy, and providing valuable insights to drive business growth.
Key Responsibilities:
- Design, develop, and maintain ETL processes to extract, transform, and load data from various sources.
- Monitor and optimize ETL workflows to ensure data quality and performance.
- Collaborate with cross-functional teams to gather and understand data requirements.
- Optimize and tune ETL processes for performance and scalability.
- Create and maintain documentation for ETL processes and data analytics solutions.
- Create and maintain data models to support reporting and analysis needs.
- Utilize expert knowledge of Go, Python, SQL, git, JSON, YAML, CSV, and MS Excel
- Working knowledge and experience with Ruby, Bash, Argo CD/Workflow, Kubernetes (K8s), containers, GitHub actions, Linux, and AWS to enhance data operations.
- Collaborate with DevOps teams to deploy ETL solutions efficiently in a Kubernetes environment using CI/CD pipelines.
- Support and troubleshoot ETL processes and resolve any issues in a timely manner.
- Perform data analysis, develop dashboards, and present actionable insights to stakeholders.
Requirements
- Bachelor's or Master's degree in Computer Science, Information Technology, or a related field.
- Minimum of 10 years of experience in ETL operations, Systems Operations, and Data Analytics.
- Expert knowledge of SQL, git, various data formats (JSON, YAML, csv), and MS Excel.
- Expert Python and Bash skills including OO techniques.
- Proficiency in Ruby, Go, and other languages is a plus.
- Familiarity with Argo CD/Workflow, Kubernetes (K8s), containers, GitHub ac0ons, Linux, and AWS is highly desirable.
- Strong problem-solving skills and attention to detail.
- Excellent communication and collaboration skills.
- Ability to work independently and as part of a team.
- Strong proficiency in SQL and experience with MySQL or similar relational databases.
- Must be able to interact with databases using raw-SQL.
- Solid understanding of data modeling concepts and techniques.
- Experience with Jaspersob or similar reporting tools is preferred.
Desired:
- Familiarity with ELK (Elasticsearch, Logstash, Kibana) or OpenSearch for advanced log and data analysis
- Familiarity with Jasper Reports and BIRT
- Familiarity with Apache Kafka for real-time data streaming and event-driven architectures
- Experience with relational databases such as PostgreSQL and MySQL for handling structured data
- Knowledge of Druid, an open-source analytics data store, and its integration into data pipelines
- Proficiency in Apache Superset for creating interactive and insightful data visualizations
Benefits
- Health Care Plan (Medical, Dental & Vision)
- Retirement Plan (401k, IRA)
- Life Insurance (Basic, Voluntary & AD&D)
- Paid Time Off (Vacation, Sick & Public Holidays)
- Short Term & Long Term Disability
- Training & Development
- Work From Home
* Salary range is an estimate based on our AI, ML, Data Science Salary Index 💰
Tags: Architecture AWS CI/CD Computer Science CSV Data analysis Data Analytics DataOps Data pipelines Data quality DevOps Elasticsearch ELK ETL Excel Git GitHub JSON Kafka Kibana Kubernetes Linux Logstash MySQL OpenSearch Open Source Pipelines PostgreSQL Python RDBMS Ruby SQL Streaming Superset
Perks/benefits: 401(k) matching Health care Insurance
More jobs like this
Explore more career opportunities
Find even more open roles below ordered by popularity of job title or skills/products/technologies used.