Sr Data Engineer (PySpark) - Hybrid
Hartford CT- Home Office
Full Time Senior-level / Expert USD 123K - 184K
The Hartford
Get business, home and car insurance from The Hartford. Choose from a broad selection of business insurance coverages and design the right solution for your company. The Hartford offers AARP members great ways to save on car and home insurance,...We’re determined to make a difference and are proud to be an insurance company that goes well beyond coverages and policies. Working here means having every opportunity to achieve your goals – and to help others accomplish theirs, too. Join our team as we help shape the future.
Join a fast-paced and talented team to deliver Data Engineering capabilities for The Hartford’s
Commercial Data Science Data Delivery. You will have an opportunity to engage in enabling well architected cloud-based data solutions for Entity Resolution using emerging technologies such as AWS and Snowflake.
This role will have a Hybrid work arrangement, with the expectation of working in an office (Hartford, CT or Charlotte, NC) 3 days a week (Tuesday through Thursday).
Responsibilities:
- Accountable as the subject matter expert and/or technical lead for a large-scale data products. Drive End-to-End solution delivery involving multiple platforms and technologies with medium to large, complexity or oversee certain parts of very large complex implementations, leveraging ELT solutions to acquire, integrate, and operationalize data.
- Partner with architects and stakeholders to influence and implement the vision of the pipeline and data product architecture while safeguarding the integrity and scalability of of the environment.
- Articulate risks and tradeoffs of technology solutions to senior leaders with translations as needed for business leaders.
- Accountable for data pipeline and product physical solution designs across teams as well as tool recommendations
- Accountable for Data Engineering Practices across all the teams involved.
- Implement and utilize leading big data methodologies (AWS, Hadoop/EMR, Spark, Kafka, Snowflake and Talend) with cloud/on premise hybrid hosting solutions, on a multi-team/product level
Knowledge, Skills, and Abilities:
Strong Technical Knowledge (Cloud data pipelines and data consumption products)
- Leader and a team player with transformation mindset.
- Ability to lead successfully in a lean, agile, and fast-paced organization, leveraging Scaled Agile principles and ways of working.
- Guides team to mature Code quality management, DataOps principles, automated testing, and environment management practices to deliver incremental customer value.
Qualifications:
- Candidates must be authorized to work in the US without company sponsorship. The company will not support the STEM OPT I-983 Training Plan endorsement for this position.
- 5+ years in large scale big data engineering experience and designing best practices in Programming, SDLC practices, Distributed systems, Data warehousing solutions SQL and NoSQL, ETL tools, CICD, Cloud Technologies (AWS/AZURE),Python/Spark, Datamesh and Datalake, Data Fabric
- 3+ years of developing and operating production workloads in cloud infrastructure (AWS, Azure, etc)
- 3+ years of operating in a technical leadership capacity for 2+ teams.
- Exposure to AWS best practices
- Knowledge of core functional components/services of AWS – compute, storage, Edge, Database, Migration and Transfer, Networking, and Governance
Certifications/Licenses (as applicable)
- Cloud certifications preferred.
Compensation
The listed annualized base pay range is primarily based on analysis of similar positions in the external market. Actual base pay could vary and may be above or below the listed range based on factors including but not limited to performance, proficiency and demonstration of competencies required for the role. The base pay is just one component of The Hartford’s total compensation package for employees. Other rewards may include short-term or annual bonuses, long-term incentives, and on-the-spot recognition. The annualized base pay range for this role is:
$123,280 - $184,920Equal Opportunity Employer/Females/Minorities/Veterans/Disability/Sexual Orientation/Gender Identity or Expression/Religion/Age
About Us | Culture & Employee Insights | Diversity, Equity and Inclusion | Benefits
Tags: Agile Architecture AWS Azure Big Data DataOps Data pipelines Data Warehousing Distributed Systems ELT Engineering ETL Hadoop Kafka NoSQL Pipelines PySpark Python SDLC Snowflake Spark SQL STEM Talend Testing
Perks/benefits: Equity / stock options Insurance
More jobs like this
Explore more career opportunities
Find even more open roles below ordered by popularity of job title or skills/products/technologies used.