Data Engineer: ETL Developer & Cloud Engineering
PA - Philadelphia, 1800 Arch St, United States
Applications have closed
Comcast
Comcast NBCUniversal creates incredible technology and entertainment that connects millions of people to the moments and experiences that matter most.Job Summary
Global Product Business Operations is a growing team that is part of the newly formed Chief of Staff Organization under the Global Product Organization (GPO). Focusing on financial operations, Bus Ops provides strategic and operational support to enable Global Product to run effectively and efficiently. Uniquely positioned, Bus Ops supports Comcast, Sky, Xumo and CTS. Our focus areas are driving financial operations, analyzing resource and investment allocations, and sharing data analytics & reporting to inform investment decisions.Job Description
This position is ineligible for visa sponsorship. To be considered for this role, you must be legally authorized to work in the United States and not require sponsorship for employment now or in the future.
This role will be responsible for aggregating and visualizing data across Comcast and Sky for the Global Product Organization through the Business Operations and Chief of Staff function that requires capturing managing, storing utilize structured and unstructured data from internal and external sources. Establishes and builds processes and structures based on business and technical requirements to channel data from multiple inputs, route appropriately and store using any combination of distributed (cloud) structures, local databases, and other applicable storage forms as required. Develops technical tools and programming that leverage artificial intelligence, machine learning and big-data techniques to cleanse, organize and transform data and to maintain, defend and update data structures and integrity on an automated basis. Creates and establishes design standards and assurance processes for software, systems and applications development to ensure compatibility and operability of data connections, flows and storage requirements. Reviews internal and external business and product requirements for data operations and activity and suggests changes and upgrades to systems and storage to accommodate ongoing needs. Work with data modelers/analysts to understand the business problems they are trying to solve then create or augment data assets to feed their analysis. Works with moderate guidance in own area of knowledge. This will be integral achieving our goals focusing around driving financial operations, analyzing resource and investment allocations, and sharing data analytics & reporting to inform investment decisions.
Requirements:
Educational background: Computer Science, Information Technology or Software Engineering
Knowledge of Big Data Platforms and Technologies
Programming Skills: Intermediate Programming Language and Scripting skills (Python – Pandas, PySpark and Java) and the ability to review and modify code
Query Languages: Intermediate knowledge with SQL, NoSQL and GraphQL
Cloud Technologies: Aggregate, transform, and clean large datasets for storage in cloud environments through AWS, Azure, Google Cloud. Intermediate knowledge of database services (EC2, S3, Redshift, Lambda, Synapse, etc.) and storage solutions
Big Data: Apache Airflow, Databricks, ETL
Database: Experience with relational (postgres/Oracle), non relational databases (MongoDB, DynamoDB)
Monitoring and Logging: Intermediate proficiency (e.g. SNMP, Prometheus, ELK, Grafana, VictoriaMetrics)
Design, develop, and maintain ETL pipelines to ingest and process data from diverse internal and external data sources
Aggregate, transform, and clean large datasets for storage in cloud environments (e.g., AWS, Azure, or GCP)
Ensure data integrity and quality across data pipelines and storage systems. Optimize data pipelines for performance, scalability, and reliability.
Establishing API connections to connect various tools and data
Identifying and standing up a new visualization tool to use for dashboards and visualizations (i.e. Tableau)
Automating, simplifying, and enhancing processes using programming languages, Excel Power Query, macros, etc.
Preferred Qualifications
Experience with containerization (Docker) or orchestration tools (Kubernetes) is a plus.
Familiarity with CI/CD for data pipelines.
Exposure to machine learning pipelines or integration with analytics platforms is a bonus.
Disclaimer: This information has been designed to indicate the general nature and level of work performed by employees in this role. It is not designed to contain or be interpreted as a comprehensive inventory of all duties, responsibilities and qualifications.
Skills
Databricks Platform, ETL Tools, Java, NoSQL, Python (Programming Language), Structured Query Language (SQL), Tableau DesktopWe believe that benefits should connect you to the support you need when it matters most, and should help you care for those who matter most. That's why we provide an array of options, expert guidance and always-on tools that are personalized to meet the needs of your reality—to help support you physically, financially and emotionally through the big milestones and in your everyday life.
Please visit the benefits summary on our careers site for more details.
Education
Bachelor's DegreeWhile possessing the stated degree is preferred, Comcast also may consider applicants who hold some combination of coursework and experience, or who have extensive related professional experience.Certifications (if applicable)
Relevant Work Experience
2-5 YearsComcast is proud to be an equal opportunity workplace. We will consider all qualified applicants for employment without regard to race, color, religion, age, sex, sexual orientation, gender identity, national origin, disability, veteran status, genetic information, or any other basis protected by applicable law.* Salary range is an estimate based on our AI, ML, Data Science Salary Index 💰
Tags: Airflow APIs AWS Azure Big Data CI/CD Computer Science Data Analytics Databricks DataOps Data pipelines Docker DynamoDB EC2 ELK Engineering ETL Excel GCP Google Cloud Grafana GraphQL Java Kubernetes Lambda Machine Learning MongoDB NoSQL Oracle Pandas Pipelines PostgreSQL PySpark Python RDBMS Redshift SQL Tableau Unstructured data
Perks/benefits: Career development
More jobs like this
Explore more career opportunities
Find even more open roles below ordered by popularity of job title or skills/products/technologies used.