Sr Data Operations and Reliability Engineer

Lakeside, United States

Texas Capital

Bank boldly with Texas Capital.

View all jobs at Texas Capital

Apply now Apply later

 

Texas Capital is built to help businesses and their leaders. Our depth of knowledge and expertise allows us to bring the best of the big firms at a scale that works for our clients, with highly experienced bankers who truly invest in people’s success — today and tomorrow.  

  

While we are rooted in core financial products, we are differentiated by our approach. Our bankers are seasoned financial experts who possess deep experience across a multitude of industries. Equally important, they bring commitment — investing the time and resources to understand our clients’ immediate needs, identify market opportunities and meet long-term objectives. At Texas Capital, we do more than build business success. We build long-lasting relationships. 

 

Texas Capital provides a variety of benefits to colleagues, including health insurance coverage, wellness program, fertility and family building aids, life and disability insurance, retirement savings plans with a generous 401K match, paid leave programs, paid holidays, and paid time off (PTO). 

 

Headquartered in Dallas with offices in Austin, Fort Worth, Houston, Richardson, Plano and San Antonio, Texas Capital was recently named Best Regional Bank in 2024 by Bankrate and was named to The Dallas Morning News’ Dallas-Fort Worth metroplex Top Workplaces 2023 and GoBankingRate’s 2023 list of Best Regional Banks. For more information about joining our team, please visit us at www.texascapitalbank.com

 

We are looking for a highly skilled and proactive Data Reliability and Operations Engineer to ensure the seamless operation of our cloud-based data pipelines and critical business processes. The ideal candidate will monitor and maintain data workflows, identify recurring issues, and implement automation to enhance efficiency. They will also collaborate with data engineering teams to optimize and scale pipelines using modern cloud technologies like AWS Cloud Services, Snowflake, PySpark, AWS Glue, AWS Airflow, and related tools.


Responsibilities

  • Monitor, troubleshoot, and maintain the performance of cloud-based data pipelines based on tools like Airflow, AWS Glue, Coalesce, and SnowFlake.
  • Ensure the smooth execution of business-critical processes and quickly address operational issues.
  • Set up real-time monitoring and alerting systems with observability tools (e.g., DataDog, CloudWatch).
  • Reliability and Problem Solving
  • Identify patterns and root causes of recurring issues in data workflows and implement long-term solutions.
  • Develop and maintain robust error-handling and data recovery mechanisms.
  • Ensure data pipelines meet SLAs for uptime, accuracy, and performance.
  • Automation and Optimization
  • Automate repetitive tasks, including data validation, pipeline deployment, and monitoring processes.
  • Able to guide teams in building scalable and self-healing systems to minimize manual intervention.
  • Collaboration and Continuous Improvement
  • Partner with data engineering teams to improve pipeline architecture, scalability, and performance.
  • Contribute to cloud migration initiatives, transitioning data workflows from legacy systems to modern platforms like Snowflake, AWS Glue, and Coalesce.
  • Promote best practices for data reliability, pipeline design, and operational efficiency.

Qualifications

  • Bachelor’s degree in Computer Science, Data Engineering, or a related field.
  • 5+ years of experience in data operations, reliability engineering, or a related role.
  • Expertise in AWS services (S3, Glue, Airflow, CloudWatch, CloudTrail, etc.) and data warehouse solutions like Snowflake.
  • Hands-on experience with PySpark, Python, and workflow orchestration tools like Apache Airflow.
  • Strong SQL skills for querying and optimizing data pipelines.
  • Knowledge of CI/CD pipelines for automated deployments.
  • Familiarity with data governance, CMDB registration, and compliance standards.
  • Experience automating and deploying database changes using tools like Liquibase.
  • Knowledge of cloud cost optimization strategies and governance frameworks.
  • Soft Skills
  • Exceptional analytical and problem-solving abilities.
  • Strong communication skills to collaborate effectively with engineering teams and stakeholders.
  • Self-motivated, proactive, and committed to continuous learning and improvement.

The duties listed above are the essential functions, or fundamental duties within the job classification.  The essential functions of individual positions within the classification may differ. Texas Capital Bank may assign reasonably related additional duties to individual employees consistent with standard departmental policy.Texas Capital is an Equal Opportunity Employer. 

Apply now Apply later

* Salary range is an estimate based on our AI, ML, Data Science Salary Index 💰

Job stats:  0  0  0

Tags: Airflow Architecture AWS AWS Glue CI/CD Classification Computer Science Data governance DataOps Data pipelines Data warehouse Engineering Pipelines PySpark Python Snowflake SQL

Perks/benefits: 401(k) matching Career development Fertility benefits Flex vacation Health care Insurance Wellness

Region: North America
Country: United States

More jobs like this