Data Test Automation Engineer (Selenium & ETL/Big Data Testing)
Bengaluru - BCIT, India
Synechron
Synechron is an innovative global consulting firm delivering industry-leading digital solutions to transform and empower businesses.Job Summary
Synechron seeks a meticulous and experienced Data Test Automation Engineer to enhance our data validation and automation efforts. The role is pivotal in verifying data quality, ensuring system integrity, and automating testing within complex data pipelines, including ETL and Big Data environments. The ideal candidate will bring expertise in Selenium automation frameworks and hands-on experience with large-scale data systems, contributing to the reliability and scalability of our data analytics initiatives.
This role enables data-driven decision-making and supports our advanced analytics and digital transformation objectives by ensuring high standards of data accuracy and system performance.
Software Requirements
Required:
- Selenium WebDriver (Java/Python bindings) — mandatory for automation scripting
- SQL and data querying tools (Oracle, SQL Server, etc.) — for validation and verification
- ETL tools (preferably Informatica, Talend, or equivalent)
- Big Data frameworks (Hadoop, Spark, or similar) — basic understanding needed
- Data validation and testing tools (e.g., data compare tools, custom scripts)
Preferred:
- Scripting languages: Python and Java for automation and scripting tasks
- Data warehousing tools such as Snowflake, Redshift, or similar
- Automation frameworks (e.g., TestNG, JUnit)
- Cloud platforms (AWS, Azure) knowledge to support data processing environments
Overall Responsibilities
- Design and execute comprehensive test plans for data validation within ETL workflows and Big Data architectures
- Develop, maintain, and execute automated test scripts using Selenium to verify data integrity across multiple systems
- Validate data accuracy, completeness, and consistency across sources, staging, and target systems
- Collaborate with data engineers and developers to understand data workflows, identify testing needs, and address quality issues
- Document defects, track resolution, and communicate findings clearly to technical and non-technical stakeholders
- Perform performance testing on data pipelines to evaluate scalability and efficiency under load
- Maintain detailed testing documentation, including test cases, scripts, defect logs, and reports
- Continuously improve testing processes by integrating best practices for automation, data governance, and compliance standards
Technical Skills (By Category)
Programming Languages:
- Required: Java, Python (both essential for automation scripting and data validation)
- Preferred: Shell scripting, Scala (useful for Big Data environments)
Databases/Data Management:
- Required: SQL proficiency for data validation tasks (querying, data comparison)
- Preferred: Experience with data warehousing and Data Lake architectures (e.g., Hadoop, Snowflake)
Cloud Technologies:
- Preferred: Familiarity with cloud platforms supporting data pipelines (AWS, Azure)
Frameworks and Libraries:
- Required: Selenium WebDriver, TestNG/JUnit for automation
- Preferred: Pytest, pandas for data processing and validation scripts
Development Tools and Methodologies:
- Required: Version control (Git), Agile/Scrum workflows
- Preferred: Continuous Integration tools (Jenkins)
Security Protocols:
- Not specific but must understand data security and governance standards relevant to testing environments
Experience Requirements
- Minimum 3-5 years of experience in data testing, especially within ETL and Big Data environments
- Proven expertise in testing data workflows, validating large datasets, and automating test cases using Selenium
- Solid understanding of data warehousing concepts and architecture (data lakes, marts, warehouses)
- Prior experience in scripting and automation using Java and Python
- Familiarity with cloud-based data platforms and tools is a plus
- Experience with data governance, compliance standards, and performance testing of data pipelines
Day-to-Day Activities
- Analyze project requirements and develop detailed test plans for data workflows
- Build and execute automated test scripts for data validation using Selenium
- Collaborate with data engineers to understand ETL and data pipeline architectures
- Conduct functional, regression, and performance testing of data pipelines
- Log defects, perform root cause analysis, and coordinate with the team for resolution
- Review test results, prepare status reports, and suggest improvements
- Support continuous integration and deployment pipelines to streamline data testing processes
- Stay updated on emerging data testing tools and industry best practices
Qualifications
- Educational Background: Bachelor’s or higher degree in Computer Science, Information Technology, or related disciplines; equivalent industry experience acceptable
- Certifications: Relevant certifications in Data Testing, Big Data, or automation frameworks (preferred)
- Training in data governance, security, and compliance standards is advantageous
- Commitment to continuous professional development in data technologies and testing methodologies
Professional Competencies
- Strong analytical and critical thinking skills for complex data validation
- Effective communication skills for engaging with cross-functional teams and stakeholders
- Ability to prioritize tasks and manage time efficiently in a fast-paced environment
- Team collaboration skills and adaptability to evolving project requirements
- Continuous learning mindset focused on emerging data tools and best practices
- Problem-solving orientation with attention to detail to identify and resolve data inconsistencies
SYNECHRON’S DIVERSITY & INCLUSION STATEMENT
Diversity & Inclusion are fundamental to our culture, and Synechron is proud to be an equal opportunity workplace and is an affirmative action employer. Our Diversity, Equity, and Inclusion (DEI) initiative ‘Same Difference’ is committed to fostering an inclusive culture – promoting equality, diversity and an environment that is respectful to all. We strongly believe that a diverse workforce helps build stronger, successful businesses as a global company. We encourage applicants from across diverse backgrounds, race, ethnicities, religion, age, marital status, gender, sexual orientations, or disabilities to apply. We empower our global workforce by offering flexible workplace arrangements, mentoring, internal mobility, learning and development programs, and more.
All employment decisions at Synechron are based on business needs, job requirements and individual qualifications, without regard to the applicant’s gender, gender identity, sexual orientation, race, ethnicity, disabled or veteran status, or any other characteristic protected by law.
* Salary range is an estimate based on our AI, ML, Data Science Salary Index 💰
Tags: Agile Architecture AWS Azure Big Data Computer Science Data Analytics Data governance Data management Data pipelines Data quality Data Warehousing ETL Git Hadoop Informatica Java Jenkins Oracle Pandas Pipelines Python Redshift Scala Scrum Security Selenium Shell scripting Snowflake Spark SQL Talend Testing
Perks/benefits: Career development
More jobs like this
Explore more career opportunities
Find even more open roles below ordered by popularity of job title or skills/products/technologies used.