Data Engineer - Pyspark,SQL

Gera Commerzone SEZ, Pune, India

Barclays

Barclays is a British universal bank. Our businesses include consumer banking, as well as a top-tier, global corporate and investment bank.

View all jobs at Barclays

Apply now Apply later

Job Description

Purpose of the role

To build and maintain the systems that collect, store, process, and analyse data, such as data pipelines, data warehouses and data lakes to ensure that all data is accurate, accessible, and secure. 

Accountabilities

  • Build and maintenance of data architectures pipelines that enable the transfer and processing of durable, complete and consistent data.
  • Design and implementation of data warehoused and data lakes that manage the appropriate data volumes and velocity and adhere to the required security measures.
  • Development of processing and analysis algorithms fit for the intended data complexity and volumes.
  • Collaboration with data scientist to build and deploy machine learning models.

Assistant Vice President Expectations

  • To advise and influence decision making, contribute to policy development and take responsibility for operational effectiveness. Collaborate closely with other functions/ business divisions.
  • Lead a team performing complex tasks, using well developed professional knowledge and skills to deliver on work that impacts the whole business function. Set objectives and coach employees in pursuit of those objectives, appraisal of performance relative to objectives and determination of reward outcomes
  • If the position has leadership responsibilities, People Leaders are expected to demonstrate a clear set of leadership behaviours to create an environment for colleagues to thrive and deliver to a consistently excellent standard. The four LEAD behaviours are: L – Listen and be authentic, E – Energise and inspire, A – Align across the enterprise, D – Develop others.
  • OR for an individual contributor, they will lead collaborative assignments and guide team members through structured assignments, identify the need for the inclusion of other areas of specialisation to complete assignments. They will identify new directions for assignments and/ or projects, identifying a combination of cross functional methodologies or practices to meet required outcomes.
  • Consult on complex issues; providing advice to People Leaders to support the resolution of escalated issues.
  • Identify ways to mitigate risk and developing new policies/procedures in support of the control and governance agenda.
  • Take ownership for managing risk and strengthening controls in relation to the work done.
  • Perform work that is closely related to that of other areas, which requires understanding of how areas coordinate and contribute to the achievement of the objectives of the organisation sub-function.
  • Collaborate with other areas of work, for business aligned support areas to keep up to speed with business activity and the business strategy.
  • Engage in complex analysis of data from multiple sources of information, internal and external sources such as procedures and practises (in other areas, teams, companies, etc).to solve problems creatively and effectively.
  • Communicate complex information. 'Complex' information could include sensitive information or information that is difficult to communicate because of its content or its audience.
  • Influence or convince stakeholders to achieve outcomes.

All colleagues will be expected to demonstrate the Barclays Values of Respect, Integrity, Service, Excellence and Stewardship – our moral compass, helping us do what we believe is right. They will also be expected to demonstrate the Barclays Mindset – to Empower, Challenge and Drive – the operating manual for how we behave.

Join us as a Data Engineer - Pyspark,SQL at Barclays, where you'll spearhead the evolution of our digital landscape, driving innovation and excellence. You'll harness cutting-edge technology to revolutionise our digital offerings, ensuring unparalleled customer experiences. As a part of team of developers, you will deliver technology stack, using strong analytical and problem solving skills to understand the business requirements and deliver quality solutions.

To be successful as a Data Engineer - Pyspark,SQL you should have experience with:

  • Hands on experience in Pyspark and strong knowledge on Dataframes, RDD and SparkSQL

  • Hands on experience in Pyspark performance optimization techniques .

  • Hands on Experience in developing, testing and maintaining applications on AWS Cloud.

  • Strong hold on AWS Data Analytics Technology Stack (Glue, S3, Lambda, Lake formation, Athena)

  • Design and implement scalable and efficient data transformation/storage solutions with open table formats such as DELTA, Iceberg, Hudi.

  • Experience in using DBT (Data Build Tool) with snowflake/Athena/Glue for ELT pipeline development.

  • Experience in Writing advanced SQL and PL SQL programs.

  • Hands On Experience for building reusable components using Snowflake and AWS Tools/Technology

  • Should have worked at least on two major project implementations.

  • Exposure to data governance or lineage tools such as Immuta and Alation is added advantage.

  • Experience in using Orchestration tools such as Apache Airflow or Snowflake Tasks is added advantage.

  • Knowledge on Ab-initio ETL tool is a plus

Some other highly valued skills includes:

  • Ability to engage with Stakeholders, elicit requirements/ user stories and translate requirements into ETL components

  • Ability to understand the infrastructure setup and be able to provide solutions either individually or working with teams.

  • Good knowledge of Data Marts and Data Warehousing concepts.

  • Resource should possess good analytical and Interpersonal skills.

  • Implement Cloud based Enterprise data warehouse with multiple data platform along with Snowflake and NoSQL environment to build data movement strategy.

You may be assessed on key critical skills relevant for success in role, such as risk and controls, change and transformation, business acumen, strategic thinking and digital and technology, as well as job-specific technical skills.

This role is based out of Pune.

Apply now Apply later

* Salary range is an estimate based on our AI, ML, Data Science Salary Index 💰

Job stats:  0  0  0
Category: Engineering Jobs

Tags: Airflow Architecture Athena AWS Data Analytics Data governance Data pipelines Data warehouse Data Warehousing dbt ELT ETL Lake Formation Lambda Machine Learning ML models NoSQL Pipelines PySpark Security Snowflake SQL Testing

Region: Asia/Pacific
Country: India

More jobs like this