Data Engineer Intern
Oakland, CA, US, 94612
Pacific Gas and Electric Company
Pacific Gas and Electric Company (PG&E) provides natural gas and electric service to residential and business customers in northern and central California.Requisition ID # 163610
Job Category: Administrative / Clerical
Job Level: Individual Contributor
Business Unit: Operations - Other
Work Type: Hybrid
Job Location: Oakland; San Ramon
Department Overview
Years of drought, extreme heat, and 129 million dead trees have created a “new normal” in California that requires PG&E to change its approach to how it designs and operates the electric system. Extreme weather events driven by climate change are causing unprecedented wildfires. The company is bolstering wildfire prevention and emergency response efforts, putting in place new and enhanced safety measures, and hardening its electric system to help reduce wildfire risks and keep customers safe. A key tool in our wildfire safety program is Public Safety Power Shutoff (PSPS), which is designed to de-energize portions of PG&E’s system during extreme weather conditions to mitigate the risk of catastrophic fire.
The Public Safety Power Shutoff team has a critical role in ensuring safe and effective execution of de-energization events. We apply critical thinking, data, and Lean principles to execute de-energization events safely and efficiently for our customers.
Position Summary
Our team is seeking a Data Engineer to drive improvements in PSPS operational readiness. This team member will work cross-functionally to deliver valuable data to our internal and external PG&E partners, data that will be critical in the safe and successful execution of PSPS events. You will expand and optimize our data and data pipeline architecture and optimize data flow and collection across multiple teams. The right candidate will be excited by the prospect of designing and developing data products that will support our company’s wildfire safety program, owning complex problem analysis and resolution, and working in a fast-paced and collaborative environment.
PG&E is providing the hourly rate range that the company in good faith believes it might pay for this position at the time of the job posting. This compensation range is specific to the locality of the job. The actual hourly rate paid to an individual will be based on multiple factors, including, but not limited to, specific skills, education, licenses or certifications, experience, market value, geographic location, and internal equity.
The hourly rate for this position ranges from $22.77 to $36.42.
Job Responsibilities
- The Data Engineer will lead and/or support activities as described below working with a cross-functional team of program and product managers, business analyst and other data engineers.
- Design, build and test high-performance data pipelines required for optimal extraction, transformation, and loading of data from a wide variety of data sources
- Assemble large and complex datasets that meet functional and non-functional business requirements
- Optimize data storage and query performance; ensure data integrity, cleanliness, and availability; and document data sources, methodologies and test plans/results.
- Design, build and test analytics, visualization and dashboards to provide actionable insights and key business metrics
- Identify, design, and implement process improvements by automating and integrating manual processes for greater efficiency and scalabilityCollaborate with stakeholders across organizations to support their data analytics needs.
- Additionally, the Data Engineer will represent the PSPS program as a subject matter expert and stakeholder on process improvement teams, regulatory filings, data requests, and various stakeholder groups.
Qualifications
Minimum Qualifications:
- BA/BS in Computer Science, Applied Math, or equivalent experience and/or field of study
- Students must be continuing their education towards their degree during and/or after the internship
Desired:
- Familiarity with Palantir platforms, specifically Foundry
- Experience developing data tools in Python using Spark
- Experience building and optimizing data pipelines and large data sets
- Experience performing root cause analysis on data and processes to answer specific business questions and identify opportunities for improvement
- Experience supporting and working with cross-functional teams in a dynamic environment
- Influence and negotiation ability, including strong meeting facilitation skills, to effectively prioritize work based on business need and risk assessments
- Ability to tackle complex and open-ended problems in unstructured environments
- Curiosity and willingness to impact beyond defined role
- Develop and maintain data pipelines using PySpark within Foundry
- Experience in databases and query languages such as SQL.
- Collaborate with engineers and analysts to understand data requirements and translate them into efficient data solutions.
- Build and deploy data visualizations and dashboards to communicate key insights to stakeholders.
- Contribute to the development and improvement of data quality monitoring and validation processes.
- Proactively identify and address potential data issues and bottlenecks.
- Document data processes and maintain clear and concise code
- Familiarity with data engineering concepts and best practices.
PG&E is unable to provide VISA sponsorship to students on an F-1, J-1 or other student visa for this position
Tags: Architecture Computer Science Data Analytics Data pipelines Data quality Engineering Mathematics Pipelines PySpark Python Spark SQL
Perks/benefits: Equity / stock options Team events
More jobs like this
Explore more career opportunities
Find even more open roles below ordered by popularity of job title or skills/products/technologies used.