Data Engineer

Issaquah, WA, US

Apply now Apply later

Costco IT is responsible for the technical future of Costco Wholesale, the third largest retailer in the world with wholesale operations in fourteen countries. Despite our size and explosive international expansion, we continue to provide a family, employee centric atmosphere in which our employees thrive and succeed.

This is an environment unlike anything in the high-tech world and the secret of Costco’s success is its culture. The value Costco puts on its employees is well documented in articles from a variety of publishers including Bloomberg and Forbes. Our employees and our members come FIRST.  Costco is well known for its generosity and community service and has won many awards for its philanthropy. The company joins with its employees to take an active role in volunteering by sponsoring many opportunities to help others. 

Come join the Costco Wholesale IT family. Costco IT is a dynamic, fast-paced environment, working through exciting transformation efforts. We are building the next generation retail environment where you will be surrounded by dedicated and highly professional employees.

Data Engineers are responsible for developing and operationalizing data pipelines to make data available for consumption (i.e. reports and advanced analytics). This includes data ingestion, data transformation, data validation/quality, data visualization, data pipeline optimization, orchestration; and engaging with DevOps Engineers during CI/CD. The Data Engineer role requires a grounding in programming and SQL, followed by expertise in data storage, modeling, cloud, data warehousing, and data lakes. This role also will partner closely with Product Owners, Data Architects, Platform/DevOps Engineers, etc. to design, build, test, and automate data pipelines that are relied upon across the company as the single source of truth.

The Data Engineer is responsible for data across the IS Sustainability stack. This is a new team at Costco and will be fast paced, highly visible, supporting the business goals by being an industry leader in this space. This role is focused on data engineering to build and deliver automated data pipelines from a variety of internal and external data sources.

If you want to be a part of one of the worldwide BEST companies “to work for”, simply apply and let your career be reimagined.

ROLE

 ●      Works in tandem with Data Architects to align on data architecture requirements provided by the requestor.

●      Defines and maintains optimal data pipeline architecture.

●      Analyzes data to spot anomalies, trends and correlate data to ensure Data Quality.

●      Develops data pipelines to store data in defined data models / structures.

●      to improve data reliability, efficiency and quality of data management

●      Performs peer review for another Data Engineer’s work

●      Develops and operationalizes data pipelines to create enterprise certified data sets that are made available for consumption (BI, Advanced analytics, APIs/Services).

●      Identifies ways to improve data reliability, efficiency, and quality of data management.

●      Works with area SMEs to design data pipelines and recommends ongoing optimization of data storage, data ingestion, data quality, and orchestration.

●      Identifies, designs, and implements internal process improvements: automating manual processes, optimizing data delivery.

●      Conducts ad-hoc data retrieval for business reports and dashboards.

●      Assesses the integrity of data from multiple sources.

●      Designs, develops, and implements ETL/ELT processes using Informatica Intelligent Cloud Services (IICS) and Azure Data Factory (ADF).

●      Uses Azure services, such as Databricks, Azure SQL DW (Synapse), Data Lake Storage, Azure Event Hub, Cosmos, Delta-Lake to improve and speed up delivery of our data products and services.

●      Develops and implements PowerBI reports and applications.

●      Implements big data and NoSQL solutions by developing scalable data processing platforms to drive high-value insights to the organization.

●      Leads the analysis by applying statistics, machine learning, and analytic approaches to predict and optimize business outcomes.

●      Designs and builds ML/DL models to solve business problems.

●      Frames a problem correctly and comes up with a hypothesis.

●      Communicates technical concepts to non-technical audiences both in written and verbal form.

REQUIRED

●      Experience engineering and operationalizing data pipelines with large and complex datasets.

●      Hands-on experience with Informatica PowerCenter and/or IICS.

●      Experience with Cribl, Confluent/Kafka, Big Panda.

●      Experience working with Cloud technologies, such as ADLS, Azure Databricks, Spark, Azure Synapse, Cosmos DB, and other big data technologies.

●      Extensive experience working with various data sources (DB2, SQL, Oracle, flat files (csv, delimited), APIs, XML, JSON).

●      Advanced SQL skills. Solid understanding of relational databases and business data; ability to write complex SQL queries against a variety of data sources.

●      Experience with Data Modeling, ETL, and Data Warehousing.

●      Strong understanding of database storage concepts (data lake, relational databases, NoSQL, Graph, data warehousing).

●      Experience in delivering business insights using advanced statistical and machine learning models and visualization.

●      Proficiency in working with diverse databases and other data sources.

●      Experience with Git / Azure DevOps.

●      Experience delivering data solutions through agile software development methodologies.

●      Demonstrates ability to communicate technical concepts to non-technical audiences both in written and verbal form.

●      Demonstrates strong understanding with coding and programming concepts to build data pipelines (e.g. data transformation, data quality, data integration, etc.).

●      Demonstrates strong understanding of database storage concepts (data lake, relational databases, NoSQL, Graph, data warehousing).

Recommended

 ●      Graduate degree in Computer Science, Data Science, and Statistics/Mathematics or related field.

●      Azure, GCP Certifications.

●      Experience implementing data integration techniques, such as event/message based integration (Kafka, Azure Event Hub), ETL.

●      Exposure to the retail industry.

●      Experience with UC4 Job Scheduler.

●      Strong proficiency in Machine Learning, Statistical and Reporting tools (Python, R, SQL, PowerBI).

●      Knowledge of Deep Learning and Neural Networks, and its applications.

●      Strong experience working in Cloud (Azure, GCP) based analytics platform.

●      Knowledge of Agile software development.

●      Experience in software development.

●      Excellent verbal and written communication skills.

●      Proficient in Google Workspace applications, including Sheets, Docs, Slides, and Gmail.

Required Documents

●      Cover Letter

●      Resume

 

California applicants, please click here to review the Costco Applicant Privacy Notice.

 

Pay Ranges: 

Level SR  - $150,000 - $190,000, Bonus and Restricted Stock Unit (RSU) eligible

We offer a comprehensive package of benefits including paid time off, health benefits - medical/dental/vision/hearing aid/pharmacy/behavioral health/employee assistance, health care reimbursement account, dependent care assistance plan, short-term disability and long-term disability insurance, AD&D insurance, life insurance, 401(k), stock purchase plan to eligible employees.

Costco is committed to a diverse and inclusive workplace. Costco is an equal opportunity employer. Qualified applicants will receive consideration for employment without regard of race, national origin, gender, gender identity, sexual orientation, protected veteran status, disability, age, or any other legally protected status. If you need assistance and/or a reasonable accommodation due to a disability during the application or the recruiting process, please send a request to IT-Recruiting@costco.com

If hired, you will be required to provide proof of authorization to work in the United States. 
 

Apply now Apply later
Job stats:  0  0  0
Category: Engineering Jobs

Tags: Agile APIs Architecture Azure Big Data CI/CD Computer Science Cosmos DB CSV Databricks Data management Data pipelines Data quality Data visualization Data Warehousing DB2 Deep Learning DevOps ELT Engineering ETL GCP Git Informatica JSON Kafka Machine Learning Mathematics ML models NoSQL Oracle Pipelines Power BI Privacy Python R RDBMS Spark SQL Statistics XML

Perks/benefits: Career development Equity / stock options Health care Insurance Salary bonus

Region: North America
Country: United States

More jobs like this