Data Engineer/Software Engineer
USA VA Herndon (13461 Sunrise Valley Drive), Estados Unidos
Full Time Senior-level / Expert Clearance required USD 100K - 190K
Parsons Corporation
Parsons is a digitally enabled solutions provider and a leader in many diversified markets with a focus on national security, defense, and global infrastructure.Job Description:
Join Us in Powering the Future of AI within Cyberspace
We are seeking a Senior Data Engineer to help build and optimize the backbone of advanced AI systems. This role offers an opportunity to work on cutting-edge AI integration projects, developing high-performance data pipelines, ensuring seamless data flow, and enabling AI-driven decision-making. If you enjoy solving complex data challenges and collaborating with top-tier data scientists and software engineers, this is the role for you.
Job Description
As a Senior Data Engineer, you will be responsible for designing, building, and optimizing scalable data pipelines that support AI-driven analytical systems. This position involves developing and managing data models, automating ETL processes, optimizing data indexing and retrieval, and working closely with AI/ML teams to structure and store data effectively. The ideal candidate will have experience with Postgres (UDFs, SQL Triggers), Spark, Ray, ElasticSearch, Python, AWS S3, and structured data formats (Parquet, Avro, JSON Schema).
This hybrid role requires candidates to be local to Maryland, Northern Virginia, or the DC Metro area for occasional on-site collaboration.
Key Responsibilities
- Data Modeling and Management – Design, develop, and maintain data models throughout their lifecycle. Collaborate with various stakeholders on requirements and their realization.
- Data Transformation and Pipelines – Transform raw data from a number of different system and automate ETL pipelines to support AI-driven analytics and machine learning applications as well as log retrieval and metric analysis
- Indexing and Retrieval – Collaborate with teams to provide data indexing and retrieval using ElasticSearch or SOLR and related technologies.
- Collaboration with AI/ML Teams – Work with an AI team to identify data to capture and store for future analysis and refinement. Support AI activities and suggest data structures and models capable of supporting their work.
- Data Governance – Ensure data quality throughout its lifecycle using Triggers, User-Defined Functions (UDFs), and Spark. Handle large data streams with different transformation requirements.
- Cloud Data Management – Leverage cloud services such as AWS S3 and other cloud-based solutions for optimizing costs and secure data storage.
Required Qualifications
- U.S. Citizenship with an active TS/SCI clearance.
- Bachelor’s degree in Software Engineering, Computer Science, Information Systems, or a related field. Four additional years of experience may be substituted in lieu of a degree.
- Minimum of five years of experience in back-end software development with a focus on data engineering.
Technical Skills
- Strong experience with SQL 2023, Postgres 15/16/17, including UDFs and Triggers.
- Expertise in ETL processes, data modeling, and data transformation.
- Proficiency in working with structured and semi-structured data formats (Parquet, Avro, JSON Schema) along with various schema description formats.
- Skilled in Python for scripting and data manipulation. Familiarity with other languages such as Java and Scala would be beneficial.
- Experience with Hadoop, Spark/PySpark and Ray for large-scale data processing.
- Hands-on expertise with ElasticSearch and/or SOLR for indexing and retrieval of massive datasets.
- Familiarity with HBase, Accumulo, Impala, AWS S3 and cloud-based storage solutions.
- JSON Schema and XML Schema design and enforcement.
Preferred Qualifications
- TS/SCI with Polygraph preferred.
- Ability to communicate highly technical analyses to non-technical audiences, including senior government representatives and executives.
- Experience with database architecture, UX, and software development/deployment tools such as Git, Artifactory, and SonarQube.
- Experience working in Agile/SAFe development environments.
- Background in modeling and processing large datasets.
Security Clearance Requirement
- Minimum Clearance Required: Top Secret
Minimum Clearance Required to Start:
Top SecretThis position is part of our Federal Solutions team.The Federal Solutions segment delivers resources to our US government customers that ensure the success of missions around the globe. Our intelligent employees drive the state of the art as they provide services and solutions in the areas of defense, security, intelligence, infrastructure, and environmental. We promote a culture of excellence and close-knit teams that take pride in delivering, protecting, and sustaining our nation's most critical assets, from Earth to cyberspace. Throughout the company, our people are anticipating what’s next to deliver the solutions our customers need now.Salary Range: $100,900.00 - $190,200.00We value our employees and want our employees to take care of their overall wellbeing, which is why we offer best-in-class benefits such as medical, dental, vision, paid time off, 401(k), life insurance, flexible work schedules, and holidays to fit your busy lifestyle!This position will be posted for a minimum of 3 days and will continue to be posted for an average of 30 days until a qualified applicant is selected or the position has been cancelled.Parsons is an equal opportunity employer, and all qualified applicants will receive consideration for employment without regard to race, color, religion, sex, national origin, disability, veteran status or any other protected status.We truly invest and care about our employee’s wellbeing and provide endless growth opportunities as the sky is the limit, so aim for the stars! Imagine next and join the Parsons quest—APPLY TODAY!Parsons is aware of fraudulent recruitment practices. To learn more about recruitment fraud and how to report it, please refer to https://www.parsons.com/fraudulent-recruitment/.
Tags: Agile Architecture Avro AWS Computer Science Data governance Data management Data pipelines Data quality Elasticsearch Engineering ETL Git Hadoop HBase Java JSON Machine Learning Parquet Pipelines PostgreSQL PySpark Python Scala Security Spark SQL UX XML
Perks/benefits: Career development Flex hours Flex vacation Health care Insurance Startup environment Team events
More jobs like this
Explore more career opportunities
Find even more open roles below ordered by popularity of job title or skills/products/technologies used.