Software Engineer II (R-17978)

Hyderabad - India

Dun & Bradstreet

Meet sales and marketing goals, navigate global supply chains, and mitigate credit risk with nearly two centuries of D&B business data and analytics expertise

View all jobs at Dun & Bradstreet

Apply now Apply later

Why We Work at Dun & BradstreetDun & Bradstreet unlocks the power of data through analytics, creating a better tomorrow. Each day, we are finding new ways to strengthen our award-winning culture and accelerate creativity, innovation and growth. Our 6,000+ global team members are passionate about what we do. We are dedicated to helping clients turn uncertainty into confidence, risk into opportunity and potential into prosperity. Bold and diverse thinkers are always welcome. Come join us! Learn more at dnb.com/careers.

Key Responsibilities:

  • Design, build, and deploy new data pipelines within our Big Data Eco-Systems using Streamsets/Talend/Informatica BDM etc. Document new/existing pipelines, Datasets.
  • Design ETL/ELT data pipelines using StreamSets, Informatica or any other ETL processing engine. Familiarity with Data Pipelines, Data Lakes and modern Data Warehousing practices (virtual data warehouse, push down analytics etc.)
  • Expert level programming skills on Python
  • Expert level programming skills on Spark
  • Cloud Based Infrastructure: GCP
  • Experience with one of the ETL Informatica, StreamSets in creation of complex parallel loads, Cluster Batch Execution and dependency creation using Jobs/Topologies/Workflows etc.,
  • Experience in SQL and conversion of SQL stored procedures into Informatica/StreamSets, Strong exposure working with web service origins/targets/processors/executors, XML/JSON Sources and Restful API’s.
  • Strong exposure working with relation databases DB2, Oracle & SQL Server including complex SQL constructs and DDL generation.
  • Exposure to Apache Airflow for scheduling jobs
  • Strong knowledge of Big data Architecture (HDFS), Cluster installation, configuration, monitoring, cluster security, cluster resources management, maintenance, and performance tuning.
  • Create POCs to enable new workloads and technical capabilities on the Platform.  
  • Work with the platform and infrastructure engineers to implement these capabilities in production.
  • Manage workloads and enable workload optimization including managing resource allocation and scheduling across multiple tenants to fulfill SLAs.
  • Participate in planning activities, Data Science and perform activities to increase platform skills.

Key Requirements:

  • Minimum 6 years of experience in ETL/ELT Technologies, preferably StreamSets/Informatica/Talend etc.
  • Minimum of 6 years hands-on experience with Big Data technologies e.g. Hadoop, Spark, Hive.
  • Minimum 3+ years of experience on Spark.
  • Minimum 3 years of experience in Cloud environments, preferably GCP.
  • Minimum 2 years of experience working in a Big Data service delivery (or equivalent) roles focusing on the following disciplines:
  • Any experience with NoSQL and Graph databases
  • Informatica or StreamSets Data integration (ETL/ELT)
  • Exposure to role and attribute based access controls
  • Hands on experience with managing solutions deployed in the Cloud, preferably on GCP.
  • Experience working in a Global company, working in a DevOps model is a plus.
All Dun & Bradstreet job postings can be found at https://www.dnb.com/about-us/careers-and-people/joblistings.html and https://jobs.lever.co/dnb. Official communication from Dun & Bradstreet will come from an email address ending in @dnb.com.
Notice to Applicants: Please be advised that this job posting page is hosted and powered by Lever. Your use of this page is subject to Lever's Privacy Notice and Cookie Policy, which governs the processing of visitor data on this platform.
Apply now Apply later

* Salary range is an estimate based on our AI, ML, Data Science Salary Index 💰

Job stats:  0  0  0
Category: Engineering Jobs

Tags: Airflow APIs Architecture Big Data Data pipelines Data warehouse Data Warehousing DB2 DDL DevOps ELT ETL GCP Hadoop HDFS Informatica JSON NoSQL Oracle Pipelines Privacy Python R Security Spark SQL Talend XML

Region: Asia/Pacific
Country: India

More jobs like this