Staff Engineer (Big Data Apache)

Gurugram, India

Applications have closed

Nagarro

A digital product engineering leader, Nagarro drives technology-led business breakthroughs for industry leaders and challengers through agility and innovation.

View all jobs at Nagarro

Company Description

👋🏼We're Nagarro.

We are a Digital Product Engineering company that is scaling in a big way! We build products, services, and experiences that inspire, excite, and delight. We work at scale across all devices and digital mediums, and our people exist everywhere in the world (18000+ experts across 36 countries, to be exact). Our work culture is dynamic and non-hierarchical. We are looking for great new colleagues. That's where you come in!

Job Description

REQUIREMENTS:

  • Experience: 7+ Years
  • Minimum of 5 years of experience in big data engineering, with at least 2 years in a lead role.
  • Proficiency in Apache Spark for data processing.
  • Strong experience with AWS Glue for ETL processes.
  • Familiarity with other big data technologies and tools.
  • Proven ability to lead and mentor a team of engineers.
  • Strong analytical and problem-solving skills.
  • Excellent verbal and written communication skills.

 

RESPONSIBILITIES:

  • Understanding the client’s business use cases and technical requirements and be able to convert them into technical design which elegantly meets the requirements.
  • Mapping decisions with requirements and be able to translate the same to developers.
  • Identifying different solutions and being able to narrow down the best option that meets the client’s requirements.
  • Defining guidelines and benchmarks for NFR considerations during project implementation
  • Writing and reviewing design document explaining overall architecture, framework, and high-level design of the application for the developers
  • Reviewing architecture and design on various aspects like extensibility, scalability, security, design patterns, user experience, NFRs, etc., and ensure that all relevant best practices are followed.
  • Developing and designing the overall solution for defined functional and non-functional requirements; and defining technologies, patterns, and frameworks to materialize it
  • Understanding and relating technology integration scenarios and applying these learnings in projects
  • Resolving issues that are raised during code/review, through exhaustive systematic analysis of the root cause, and being able to justify the decision taken.
  • Carrying out POCs to make sure that suggested design/technologies meet the requirements.

Qualifications

Bachelor’s or master’s degree in computer science, Information Technology, or a related field.

* Salary range is an estimate based on our AI, ML, Data Science Salary Index 💰

Job stats:  1  1  0

Tags: Architecture AWS AWS Glue Big Data Computer Science Engineering ETL Security Spark

Region: Asia/Pacific
Country: India

More jobs like this