Data Test Automation Engineer (Selenium & ETL/Big Data Testing)

Bengaluru - BCIT, India

Synechron

Synechron is an innovative global consulting firm delivering industry-leading digital solutions to transform and empower businesses.

View all jobs at Synechron

Apply now Apply later

Job Summary

Synechron seeks a meticulous and experienced Data Test Automation Engineer  to enhance our data validation and automation efforts. The role is pivotal in verifying data quality, ensuring system integrity, and automating testing within complex data pipelines, including ETL and Big Data environments. The ideal candidate will bring expertise in Selenium automation frameworks and hands-on experience with large-scale data systems, contributing to the reliability and scalability of our data analytics initiatives.

This role enables data-driven decision-making and supports our advanced analytics and digital transformation objectives by ensuring high standards of data accuracy and system performance.

Software Requirements

Required:

  • Selenium WebDriver (Java/Python bindings) — mandatory for automation scripting
  • SQL and data querying tools (Oracle, SQL Server, etc.) — for validation and verification
  • ETL tools (preferably Informatica, Talend, or equivalent)
  • Big Data frameworks (Hadoop, Spark, or similar) — basic understanding needed
  • Data validation and testing tools (e.g., data compare tools, custom scripts)

Preferred:

  • Scripting languages: Python and Java for automation and scripting tasks
  • Data warehousing tools such as Snowflake, Redshift, or similar
  • Automation frameworks (e.g., TestNG, JUnit)
  • Cloud platforms (AWS, Azure) knowledge to support data processing environments

Overall Responsibilities

  • Design and execute comprehensive test plans for data validation within ETL workflows and Big Data architectures
  • Develop, maintain, and execute automated test scripts using Selenium to verify data integrity across multiple systems
  • Validate data accuracy, completeness, and consistency across sources, staging, and target systems
  • Collaborate with data engineers and developers to understand data workflows, identify testing needs, and address quality issues
  • Document defects, track resolution, and communicate findings clearly to technical and non-technical stakeholders
  • Perform performance testing on data pipelines to evaluate scalability and efficiency under load
  • Maintain detailed testing documentation, including test cases, scripts, defect logs, and reports
  • Continuously improve testing processes by integrating best practices for automation, data governance, and compliance standards

Technical Skills (By Category)

Programming Languages:

  • Required: Java, Python (both essential for automation scripting and data validation)
  • Preferred: Shell scripting, Scala (useful for Big Data environments)

Databases/Data Management:

  • Required: SQL proficiency for data validation tasks (querying, data comparison)
  • Preferred: Experience with data warehousing and Data Lake architectures (e.g., Hadoop, Snowflake)

Cloud Technologies:

  • Preferred: Familiarity with cloud platforms supporting data pipelines (AWS, Azure)

Frameworks and Libraries:

  • Required: Selenium WebDriver, TestNG/JUnit for automation
  • Preferred: Pytest, pandas for data processing and validation scripts

Development Tools and Methodologies:

  • Required: Version control (Git), Agile/Scrum workflows
  • Preferred: Continuous Integration tools (Jenkins)

Security Protocols:

  • Not specific but must understand data security and governance standards relevant to testing environments

Experience Requirements

  • Minimum 3-5 years of experience in data testing, especially within ETL and Big Data environments
  • Proven expertise in testing data workflows, validating large datasets, and automating test cases using Selenium
  • Solid understanding of data warehousing concepts and architecture (data lakes, marts, warehouses)
  • Prior experience in scripting and automation using Java and Python
  • Familiarity with cloud-based data platforms and tools is a plus
  • Experience with data governance, compliance standards, and performance testing of data pipelines

Day-to-Day Activities

  • Analyze project requirements and develop detailed test plans for data workflows
  • Build and execute automated test scripts for data validation using Selenium
  • Collaborate with data engineers to understand ETL and data pipeline architectures
  • Conduct functional, regression, and performance testing of data pipelines
  • Log defects, perform root cause analysis, and coordinate with the team for resolution
  • Review test results, prepare status reports, and suggest improvements
  • Support continuous integration and deployment pipelines to streamline data testing processes
  • Stay updated on emerging data testing tools and industry best practices

Qualifications

  • Educational Background: Bachelor’s or higher degree in Computer Science, Information Technology, or related disciplines; equivalent industry experience acceptable
  • Certifications: Relevant certifications in Data Testing, Big Data, or automation frameworks (preferred)
  • Training in data governance, security, and compliance standards is advantageous
  • Commitment to continuous professional development in data technologies and testing methodologies

Professional Competencies

  • Strong analytical and critical thinking skills for complex data validation
  • Effective communication skills for engaging with cross-functional teams and stakeholders
  • Ability to prioritize tasks and manage time efficiently in a fast-paced environment
  • Team collaboration skills and adaptability to evolving project requirements
  • Continuous learning mindset focused on emerging data tools and best practices
  • Problem-solving orientation with attention to detail to identify and resolve data inconsistencies

S​YNECHRON’S DIVERSITY & INCLUSION STATEMENT
 

Diversity & Inclusion are fundamental to our culture, and Synechron is proud to be an equal opportunity workplace and is an affirmative action employer. Our Diversity, Equity, and Inclusion (DEI) initiative ‘Same Difference’ is committed to fostering an inclusive culture – promoting equality, diversity and an environment that is respectful to all. We strongly believe that a diverse workforce helps build stronger, successful businesses as a global company. We encourage applicants from across diverse backgrounds, race, ethnicities, religion, age, marital status, gender, sexual orientations, or disabilities to apply. We empower our global workforce by offering flexible workplace arrangements, mentoring, internal mobility, learning and development programs, and more.


All employment decisions at Synechron are based on business needs, job requirements and individual qualifications, without regard to the applicant’s gender, gender identity, sexual orientation, race, ethnicity, disabled or veteran status, or any other characteristic protected by law.

Candidate Application Notice

Apply now Apply later

* Salary range is an estimate based on our AI, ML, Data Science Salary Index 💰

Job stats:  0  0  0

Tags: Agile Architecture AWS Azure Big Data Computer Science Data Analytics Data governance Data management Data pipelines Data quality Data Warehousing ETL Git Hadoop Informatica Java Jenkins Oracle Pandas Pipelines Python Redshift Scala Scrum Security Selenium Shell scripting Snowflake Spark SQL Talend Testing

Perks/benefits: Career development

Region: Asia/Pacific
Country: India

More jobs like this