Data Engineer
London, Greater London, England
Snowplow Analytics
Own and unlock the value of your customer behavioral data to fuel AI-driven marketing, digital products and services, customer experiences, and fraud mitigation.
Data Engineer
London/Hybrid - Europe/Remote
#LI-Remote
About Snowplow:
Snowplow is the global leader in customer data infrastructure for AI, enabling every organization to transform raw behavioral data into governed, high-fidelity fuel for AI-powered applications—including advanced analytics, real-time personalization engines, and AI agents.
Digital-first companies like Strava, HelloFresh, Auto Trader, Burberry, and DPG Media use Snowplow to collect and process event-level data in real time, delivering it securely to their warehouse, lake, or stream, and integrate deep customer context into their applications.
Thousands of companies rely on Snowplow to uncover customer insights, predict customer behaviors, hyper-personalize customer experiences, and detect fraud in real time.
The Opportunity:
You’ll be joining the AI team at an exciting moment, as our team is leading the charge in building Snowplow Signals — our new solution for enabling real-time, AI-powered customer experiences. Your work will contribute to the infrastructure that powers these capabilities, unlocking value from behavioral data in real time.
Alongside this, you’ll work on the foundational components that continue to make Snowplow powerful and flexible for data teams: developing and maintaining dbt packages that help customers accelerate their time to insight.
This is a role for someone who enjoys bridging data collection, transformation through SQL-based models, and high-level intelligence — enabling real-time pipelines that support analytics, personalization, and emerging agentic experiences.
What you’ll be doing:
● Developing and maintaining a suite of production SQL data models used by our customer base and the community. Primarily focusing on dbt for packaging and execution.● Building out our offering around data modeling. You won’t just work on the data models themselves - you’ll work closely with Product and the wider Engineering team to shape the way we collect data via our trackers to build better data models, and drive what data model tooling we provide as part of our commercial offering as well.● Building and optimizing real-time and batch data pipelines that power personalization, recommendation engines, and predictive models — from streaming ingestion and transformation using Benthos, to materializing features for agentic use cases and customer-facing AI systems. ● Supporting our prospect/customer facing teams by showcasing the possibilities with Snowplow data, such as powering personalisation and recommendation systems, or developing advanced models (marketing attribution, lifetime value, etc.). ● Being an active part in decision making on what we build to help our customers get more value out of Snowplow, and how we deliver it. You’ll bring a different perspective and we’ll want your input! ● Responsible for developing and productizing data models, focusing on scalability, performance and maintainability, and developing in-depth understanding of cloud data warehouses and common web and mobile analytics use cases.
We’d love to hear from you if:
● SQL and DBT are your thing. You master SQL, including manipulation of large data sets, performance tuning, etc ● You're adept with both batch and streaming data processing. You have experience building streaming pipelines using tools like Benthos, enabling real-time data ingestion, transformation, and delivery across various systems.● You understand feature engineering and management. You're familiar with tools like Feast for defining, materializing, and serving features in both real-time and batch contexts.● You have extensive experience using Python which is used for auto generating data models.● You are not new to engineering. You use CI/CD, and Git source control as part of your daily job. You have experience with testing frameworks.● You are a proactive learner. Eager to expand on your software engineering knowledge and adapt to new technologies essential for automating models and advancing our engineering practices.● You’re familiar with cloud technologies. You have experience working with data in a cloud data warehouse (Redshift, Snowflake, Databricks, or BigQuery)● Experience with a modern data modeling technology (DBT)● You document and communicate clearly. Some experience with technical content writing would be a plus● You are excited to work autonomously. You are able to drive technical initiatives from discovery until to the delivery phase● You know we can’t do everything today. You’ll be pragmatic and balance our speed of delivery with our commitment to providing a reliable and trusted service to customers● You want to join a remote team that depends on expert collaborators to work effectively. You’ll be a great communicator and enjoy working closely with the team
Snowplow is dedicated to building and supporting a brilliant, diverse and hugely inclusive team. We don't discriminate against gender, race, religion or belief, disability, age, marital status or sexual orientation. Whatever your background may be, we welcome anyone with talent, drive and emotional intelligence.
London/Hybrid - Europe/Remote
#LI-Remote
About Snowplow:
Snowplow is the global leader in customer data infrastructure for AI, enabling every organization to transform raw behavioral data into governed, high-fidelity fuel for AI-powered applications—including advanced analytics, real-time personalization engines, and AI agents.
Digital-first companies like Strava, HelloFresh, Auto Trader, Burberry, and DPG Media use Snowplow to collect and process event-level data in real time, delivering it securely to their warehouse, lake, or stream, and integrate deep customer context into their applications.
Thousands of companies rely on Snowplow to uncover customer insights, predict customer behaviors, hyper-personalize customer experiences, and detect fraud in real time.
The Opportunity:
You’ll be joining the AI team at an exciting moment, as our team is leading the charge in building Snowplow Signals — our new solution for enabling real-time, AI-powered customer experiences. Your work will contribute to the infrastructure that powers these capabilities, unlocking value from behavioral data in real time.
Alongside this, you’ll work on the foundational components that continue to make Snowplow powerful and flexible for data teams: developing and maintaining dbt packages that help customers accelerate their time to insight.
This is a role for someone who enjoys bridging data collection, transformation through SQL-based models, and high-level intelligence — enabling real-time pipelines that support analytics, personalization, and emerging agentic experiences.
What you’ll be doing:
● Developing and maintaining a suite of production SQL data models used by our customer base and the community. Primarily focusing on dbt for packaging and execution.● Building out our offering around data modeling. You won’t just work on the data models themselves - you’ll work closely with Product and the wider Engineering team to shape the way we collect data via our trackers to build better data models, and drive what data model tooling we provide as part of our commercial offering as well.● Building and optimizing real-time and batch data pipelines that power personalization, recommendation engines, and predictive models — from streaming ingestion and transformation using Benthos, to materializing features for agentic use cases and customer-facing AI systems. ● Supporting our prospect/customer facing teams by showcasing the possibilities with Snowplow data, such as powering personalisation and recommendation systems, or developing advanced models (marketing attribution, lifetime value, etc.). ● Being an active part in decision making on what we build to help our customers get more value out of Snowplow, and how we deliver it. You’ll bring a different perspective and we’ll want your input! ● Responsible for developing and productizing data models, focusing on scalability, performance and maintainability, and developing in-depth understanding of cloud data warehouses and common web and mobile analytics use cases.
We’d love to hear from you if:
● SQL and DBT are your thing. You master SQL, including manipulation of large data sets, performance tuning, etc ● You're adept with both batch and streaming data processing. You have experience building streaming pipelines using tools like Benthos, enabling real-time data ingestion, transformation, and delivery across various systems.● You understand feature engineering and management. You're familiar with tools like Feast for defining, materializing, and serving features in both real-time and batch contexts.● You have extensive experience using Python which is used for auto generating data models.● You are not new to engineering. You use CI/CD, and Git source control as part of your daily job. You have experience with testing frameworks.● You are a proactive learner. Eager to expand on your software engineering knowledge and adapt to new technologies essential for automating models and advancing our engineering practices.● You’re familiar with cloud technologies. You have experience working with data in a cloud data warehouse (Redshift, Snowflake, Databricks, or BigQuery)● Experience with a modern data modeling technology (DBT)● You document and communicate clearly. Some experience with technical content writing would be a plus● You are excited to work autonomously. You are able to drive technical initiatives from discovery until to the delivery phase● You know we can’t do everything today. You’ll be pragmatic and balance our speed of delivery with our commitment to providing a reliable and trusted service to customers● You want to join a remote team that depends on expert collaborators to work effectively. You’ll be a great communicator and enjoy working closely with the team
Snowplow is dedicated to building and supporting a brilliant, diverse and hugely inclusive team. We don't discriminate against gender, race, religion or belief, disability, age, marital status or sexual orientation. Whatever your background may be, we welcome anyone with talent, drive and emotional intelligence.
* Salary range is an estimate based on our AI, ML, Data Science Salary Index 💰
Job stats:
0
0
0
Category:
Engineering Jobs
Tags: BigQuery CI/CD Databricks Data pipelines Data warehouse dbt Engineering Feature engineering Git Pipelines Python Redshift Snowflake SQL Streaming Testing
Perks/benefits: Flex hours
Regions:
Remote/Anywhere
Europe
Country:
United Kingdom
More jobs like this
Explore more career opportunities
Find even more open roles below ordered by popularity of job title or skills/products/technologies used.
BI Developer jobsData Engineer II jobsStaff Data Scientist jobsPrincipal Data Engineer jobsSr. Data Engineer jobsPrincipal Software Engineer jobsStaff Machine Learning Engineer jobsData Science Manager jobsData Manager jobsData Science Intern jobsSoftware Engineer II jobsDevOps Engineer jobsBusiness Intelligence Analyst jobsJunior Data Analyst jobsData Analyst Intern jobsData Specialist jobsBusiness Data Analyst jobsLead Data Analyst jobsStaff Software Engineer jobsSr. Data Scientist jobsSenior Backend Engineer jobsData Governance Analyst jobsAI/ML Engineer jobsData Engineer III jobsResearch Scientist jobs
Consulting jobsAirflow jobsMLOps jobsOpen Source jobsKPIs jobsEconomics jobsJavaScript jobsLinux jobsKafka jobsTerraform jobsNoSQL jobsData Warehousing jobsGoogle Cloud jobsRDBMS jobsComputer Vision jobsGitHub jobsPostgreSQL jobsScikit-learn jobsR&D jobsPhysics jobsStreaming jobsData warehouse jobsBanking jobsHadoop jobsdbt jobs
Scala jobsLooker jobsClassification jobsPandas jobsBigQuery jobsOracle jobsRAG jobsReact jobsCX jobsScrum jobsPySpark jobsPrompt engineering jobsDistributed Systems jobsIndustrial jobsELT jobsJira jobsGPT jobsRedshift jobsMicroservices jobsRobotics jobsLangChain jobsTypeScript jobsSAS jobsOpenAI jobsJenkins jobs