Sr. Site Reliability Engineering - Hadoop, Spark, Hive, Python/PySpark(Platform Support)

Bangalore, India

Visa

Das digitale und mobile Zahlungsnetzwerk von Visa steht an der Spitze der neuen Zahlungstechnologien für die neue Zahlung, elektronische und kontaktlose Zahlung, die die Welt des Geldes bilden

View all jobs at Visa

Apply now Apply later

Company Description

Visa is a world leader in payments and technology, with over 259 billion payments transactions flowing safely between consumers, merchants, financial institutions, and government entities in more than 200 countries and territories each year. Our mission is to connect the world through the most innovative, convenient, reliable, and secure payments network, enabling individuals, businesses, and economies to thrive while driven by a common purpose – to uplift everyone, everywhere by being the best way to pay and be paid.

Make an impact with a purpose-driven industry leader. Join us today and experience Life at Visa.

Job Description

A Senior Site Reliability Engineer must perform a variety of tasks and demonstrate a profound understanding of Hadoop and its related tools, such as Hive, Spark, and HDFS and need to have hands on experience on PySpark, Pythin.

 

The primary responsibilities include:

•       Single Window Support: Utilize an in-depth understanding of Hadoop and its related tools, especially Hive, Spark, and HDFS, and also experience on PySpark, Python to conduct comprehensive root cause analyses, whether they are platform, data, or user code related.

•       System Configuration: Recommend necessary system changes to the DAP platform engineering team by examining system activity and user logs for triaging and troubleshooting.

•       Performance Tuning: Guide team members in crafting efficient queries by leveraging expertise in performance tuning and optimization strategies for big data technologies.

•       Issue Resolution Across Tech Teams: Troubleshoot and resolve complex technical issues. Identify root causes, determine which tech/data platform team can rectify it, and coordinate amongst those teams.

•       Reliability Engineering: Create reports to define performance and resolution metrics for proactively identifying issues and generating alerts.

•       Office Hours and Liaising: Participate in calls across different regions in multiple time zones to ensure timely client delivery.

•       Knowledge Cataloging and Sharing: Share knowledge and cross-train peers across geographic regions using wikis and communication tools. Provide communication around issues/outages affecting multiple users.

•       Develop Standards: The team should prepare standard configurations for a variety of VCA workloads to ensure jobs run with optimal settings, maintaining good cluster health while executing jobs efficiently.

•       Continuous Learning of VCA Workload: Continuously learn and stay updated with the changing nature of data science jobs to help improve cluster utilization. Additionally, with active engagement, collaboration, effective communication, quality, integrity, and reliable delivery, develop and maintain a trusted and valued relationship with the team, customers, and business partners.

This is a hybrid position. Hybrid employees can alternate time between both remote and office. Employees in hybrid roles are expected to work from the office 2-3 set days a week (determined by leadership/site), with a general guidepost of being in the office 50% or more of the time based on business needs.

Qualifications

Basic Qualifications:
2+ years of relevant work experience and a Bachelors degree, OR 5+ years of relevant work experience
Preferred Qualifications:
• Practical experience as a Hadoop system engineer, specifically in managing Hadoop platforms.
• An ability to solve intricate production problems and debug code.
• A strong understanding of data pipelines built using PySpark, Hive, and Airflow.
• Experience with scheduling tools (such as Airflow, Oozie) or in building data processing orchestration workflows.
• Proficiency in tuning the performance of applications on Hadoop platforms.
• Good knowledge of the Hadoop ecosystem, including Zookeeper, HDFS, Yarn, Hive, and Spark.
• Hands-on experience in debugging Hadoop issues, both on the platform and applications.
• An understanding of Linux, networking, CPU, memory, and storage.
• Knowledge or experience in Python.
• Excellent written and verbal communication skills.
• A strong work ethic, the ability to work quickly and smartly, and a capacity for understanding complex concepts and functionalities.

Additional Information

Visa is an EEO Employer. Qualified applicants will receive consideration for employment without regard to race, color, religion, sex, national origin, sexual orientation, gender identity, disability or protected veteran status. Visa will also consider for employment qualified applicants with criminal histories in a manner consistent with EEOC guidelines and applicable local law.

Apply now Apply later
  • Share this job via
  • 𝕏
  • or

* Salary range is an estimate based on our AI, ML, Data Science Salary Index 💰

Job stats:  0  0  0

Tags: Airflow Big Data Data pipelines Engineering Hadoop HDFS Linux Oozie Pipelines PySpark Python Spark

Region: Asia/Pacific
Country: India

More jobs like this