DataLake Administrator / Architect

Remote, United States

DMI

DMI provides enterprise-grade, intelligent digital transformation services that simplify and scale technology to meet you where you are and support you as you grow.

View all jobs at DMI

Apply now Apply later

About DMI

DMI is a leading provider of digital services and technology solutions, headquartered in Tysons Corner, VA. With a focus on end-to-end managed IT services, including managed mobility, cloud, cybersecurity, network operations, and application development, DMI supports public sector agencies and commercial enterprises around the globe. Recognized as a Top Workplace, DMI is committed to delivering secure, efficient, and cost-effective solutions that drive measurable results. Learn more at www.dminc.com

About the Opportunity

DMI, LLC is seeking a DataLake Administrator/Architect to join us. 

 

Duties and Responsibilities:

  • Manage and maintain Data Lake clusters infrastructure on premise and in cloud: installation, configuration, performance tuning and monitoring of Hadoop clusters
  • Should demonstrate strong concepts in Unix/Linux, Windows OS, cloud platforms (AWS, GCP), Kubernetes, Open Shift & Docker
  • Must have good exposure to Cloudera manager, Cloudera Navigator or similar cluster management tool
  • Collaborate and assist developers in successful implementation of their code, monitor and fine tune their process for optimum resource utilization on cluster, ability to automate run time process  
  • Must have good knowledge of HDFS, Ranger/Sentry, Hive, Impala, Spark, HBase, Kudu, Kafka, Kafka Connect,  Schema Registry, Ni-Fi, Sqoop and other Hadoop related services
  • Exposure to Data Science collaborative tools such as data science workbench, CML, anaconda, etc.  
  • Strong Networking concepts: topology, proxy, F5, firewall
  • Strong security concepts: Active directory ,Kerberos, LDAP, SAML, SSL, data encryption @rest
  • Programming language concepts: Java, Perl, python, PySpark and Unix shell scripting
  • Possess experience in cluster management, perform cluster upgrade, migration, and testing
  • Perform periodic updates to cluster and keeping the stack current
  • Ability to expand clusters by adding new nodes and rebalance cluster storage systems
  • Manage application databases, application integration, users, roles, permissions within cluster
  • Collaborate with OpenShift, Unix, network, database and security teams on cluster related matters
  • Must monitor cluster for maximum uptime, ability to research on cluster issues via logs and collaborate with support in a proactive way

Qualifications

Education and Years of Experience:  

  • Minimum10 years’ experience in advanced technologies including a minimum of 5+ years as data lake admin/architect
  • BS degree, preferably in Computer science or equivalent
  • Good communication skills with right attitude to blend in with team
  • Minimum 5 years work experience in Hadoop ecosystems (Horton HDP or Cloudera’s CDP) 

Required and Desired Skills/Certifications: 

  • Solid experience in Cloudera data lake environments both on prem and cloud
  • Solid experience in administration and set up including security topics related to a data lake
  • Strong experience architecting and designing solutions for new business needs
  • Thorough understanding and hands-on experience with implementing robust logging and tracing implementation for end to end systems traceability
  • Familiarity with Cloudera’s BDR tool to perform and monitor backups of critical data and able to restore data when in need
  • Willing and ready to get hands on code development with dev team for developing and troubleshooting, doing quick proof of concepts for exploring new solutions, products etc.
  • Experienced in working with technical teams to discuss, analyze, understand and negotiate business requirements, being able to explain to architects about the technical considerations and associated implications on the user journey/experience/requirements.
  • Experience in tuning and optimizing Hadoop environment in keeping clusters healthy and available for end users and applications with maximum cluster uptime as defined in SLA
  • Deep knowledge and related experience with Hadoop and its ecosystem components i.e. HDFS, Yarn, Hive, MapReduce, Pig, Sqoop, Oozie, Kafka, Spark, Presto and other Hadoop components

Min Citizenship Status Required: No Restrictions

 

Physical Requirements: No Physical requirement needed for this position.

Location: Remote, US

Working at DMI

DMI is a diverse, prosperous, and rewarding place to work. Being part of the DMI family means we care about your wellbeing. As such, we offer a variety of perks and benefits that help meet various interests and needs, while still having the opportunity to work directly with a number of our award-winning, Fortune 1000 clients. The following categories make up your DMI wellbeing:

  • Convenience/Concierge - Virtual visits through health insurance, pet insurance, commuter benefits, discount tickets for movies, travel, and many other items to provide convenience.
  • Development – Annual performance management, continuing education, and tuition assistance, internal job opportunities along with career enrichment and advancement to help each employee with their professional and personal development.
  • Financial – Generous 401k matches both pre-tax and post-tax (ROTH) contributions along with financial wellness education, EAP, Life Insurance and Disability help provide financial stability for each DMI employee.
  • Recognition – Great achievements do not go unnoticed by DMI through Annual Awards ceremony, service anniversaries, peer-to-peer acknowledgment, employee referral bonuses.
  • Wellness – Healthcare benefits, Wellness programs, Flu Shots, Biometric screenings, and several other wellness options.

 

Employees are valued for their talents and contributions. We all take pride in helping our customers achieve their goals, which in turn contributes to the overall success of the company.

 

 

***************** No Agencies Please *****************

 

Applicants selected may be subject to a government security investigation and must meet eligibility requirements for access to classified information. US citizenship may be required for some positions.

Apply now Apply later

* Salary range is an estimate based on our AI, ML, Data Science Salary Index 💰

Job stats:  0  0  0
Category: Architecture Jobs

Tags: Anaconda AWS Computer Science Docker GCP Hadoop HBase HDFS Java Kafka Kubernetes Linux Oozie Perl PySpark Python Research Security Shell scripting Spark Testing

Perks/benefits: Career development Health care Insurance Wellness

Regions: Remote/Anywhere North America
Country: United States

More jobs like this