Senior Data Engineer - Databricks

McLean, VA, United States

Apply now Apply later

Overview

In today’s rapidly evolving technology landscape, an organization’s data has never been a more important aspect in achieving mission and business goals. Our data exploitation experts work with our clients to support their mission and business goals by creating and executing a comprehensive data strategy using the best technology and techniques, given the challenge. 

 

At Steampunk, our goal is to build and execute a data strategy for our clients to coordinate data collection and generation, to align the organization and its data assets in support of the mission, and ultimately to realize mission goals with the strongest effectiveness possible. 

 

For our clients, data is a strategic asset. They are looking to become a facts-based, data-driven, customer-focused organization.  To help realize this goal, they are leveraging visual analytics platforms to analyze, visualize, and share information.  At Steampunk you will design and develop solutions to high-impact, complex data problems, working with the best and data practitioners around. Our data exploitation approach is tightly integrated with Human-Centered Design and DevSecOps. 

Contributions

We are looking for seasoned Senior Data Engineer to work with our team and our clients to develop enterprise grade data platforms, services, and pipelines in Databricks. We are looking for more than just a "Senior Data Engineer", but a technologist with excellent communication and customer service skills and a passion for data and problem solving. 

 

  • Lead and architect migrations of data using Databricks with focus on performance, reliability, and scalability. 
  • Assess and understand ETL jobs, workflows, data marts, BI tools, and reports 
  • Address technical inquiries concerning customization, integration, enterprise architecture and general feature/functionality of data products 
  • Experience working with database/data warehouse/data mart solutions in cloud (Preferably AWS. Alternatively Azure, GCP). 
  • Key must have skill sets – Databricks, SQL, PySpark/Python, AWS 
  • Support an Agile software development lifecycle 
  • You will contribute to the growth of our AI & Data Exploitation Practice! 

Qualifications

Required:

 

  • Ability to hold a position of public trust with the US government. 
  • 5-7 years industry experience coding commercial software and a passion for solving complex problems.  
  • 5-7 years direct experience in Data Engineering with experience in tools such as: 
  • Big data tools: Databricks, Apache Spark, Delta Lake, etc. 
  • Relational SQL (Preferably T-SQL. Alternatively pgSQL, MySQL). 
  • Data pipeline and workflow management tools: Databricks Workflows, Airflow, Step Functions, etc. 
  • AWS cloud services: Databricks on AWS, S3, EC2, RDS (or Azure equivalents). 
  • Object-oriented/object function scripting languages: PySpark/Python, Java, C++, Scala, etc. 
  • Experience working with Data Lakehouse architecture and Delta Lake/Apache Iceberg 
  • Advanced working SQL knowledge and experience working with relational databases, query authoring and optimization (SQL) as well as working familiarity with a variety of databases. 
  • Experience manipulating, processing, and extracting value from large, disconnected datasets. 
  • Ability to inspect existing data pipelines, discern their purpose and functionality, and re-implement them efficiently in Databricks. 
  • Experience manipulating structured and unstructured data. 
  • Experience architecting data systems (transactional and warehouses). 
  • Experience the SDLC, CI/CD, and operating in dev/test/prod environments. 
  • Experience with data cataloging tools such as Informatica EDC, Unity Catalog, Collibra, Alation, Purview, or DataZone is a plus. 
  • Commitment to data governance. 
  • Experience working in an Agile environment. 
  • Experience supporting project teams of developers and data scientists who build web-based interfaces, dashboards, reports, and analytics/machine learning models

About steampunk

Steampunk relies on several factors to determine salary, including but not limited to geographic location, contractual requirements, education, knowledge, skills, competencies, and experience. The projected compensation range for this position is $140,000 to $180,000.  The estimate displayed represents a typical annual salary range for this position. Annual salary is just one aspect of Steampunk’s total compensation package for employees. Learn more about additional Steampunk benefits here. 

 

Steampunk is a Change Agent in the Federal contracting industry, bringing new thinking to clients in the Homeland, Federal Civilian, Health and DoD sectors.  Through our Human-Centered delivery methodology, we are fundamentally changing the expectations our Federal clients have for true shared accountability in solving their toughest mission challenges.  As an employee owned company, we focus on investing in our employees to enable them to do the greatest work of their careers – and rewarding them for outstanding contributions to our growth. If you want to learn more about our story, visit http://www.steampunk.com. 

 

We are an equal opportunity employer and all qualified applicants will receive consideration for employment without regard to race, color, religion, sex, national origin, disability status, protected veteran status, or any other characteristic protected by law. Steampunk participates in the E-Verify program.  

Apply now Apply later
Job stats:  0  0  0
Category: Engineering Jobs

Tags: Agile Airflow Architecture AWS Azure Big Data CI/CD Databricks Data governance Data pipelines Data strategy Data warehouse EC2 Engineering ETL GCP Informatica Java Machine Learning ML models MySQL Pipelines PySpark Python RDBMS Scala SDLC Spark SQL Step Functions T-SQL Unstructured data

Perks/benefits: Career development Health care Startup environment

Region: North America
Country: United States

More jobs like this