Senior Data Engineer

Remote

Apply now Apply later

Biconomy empowers Web3 developers to build seamless, user-friendly dApps that work effortlessly across multiple blockchains. Our battle-tested modular account and execution stack eliminates traditional UX friction points, helping projects accelerate user adoption while reducing development costs. By processing over 70 million transactions across the 300+ dApps we've served, we're powering the future of onchain economies.


The Role – Powering the Next Generation of Real-Time Market Intelligence

Hypersignal is Biconomy’s high-throughput analytics engine, designed to extract millisecond-level insights from both on-chain activity and centralized exchange (CEX) trading flows. Our mission is to equip traders and protocols with alpha-rich signals across spot and perpetual futures markets.

As a Data Engineer, you’ll play a foundational role in building and scaling the infrastructure that powers Hypersignal. From ingesting terabytes of blockchain and CEX data to delivering analysis-ready datasets for our quant and product teams, you'll help create the data backbone of a truly cutting-edge trading intelligence system.


What You Will Be Doing

  • Design, build, and maintain streaming and batch ETL pipelines for on-chain sources across EVM, Solana, Sui, Starknet, and more

  • Develop NLP and sentiment pipelines for off-chain sources (Binance, Bybit, social platforms) to extract actionable market signals

  • Normalize and unify disparate market data schemas (order books, trades, liquidations, funding rates) into a single analytics model for perpetuals

  • Implement low-latency ingestion systems using Kafka, Kinesis, PubSub, WebSockets, or Firehose, with exactly-once guarantees

  • Build and optimize lakehouse/warehouse layers (Iceberg, Delta, Snowflake, BigQuery) with Z-ordering, partitioning, and materialized views

  • Enforce data quality and observability using dbt tests, Great Expectations, and OpenTelemetry

  • Collaborate with quants and backend engineers to deliver data via GraphQL/REST APIs and feature stores

  • Continuously optimize performance, cost, and scalability across AWS/GCP infrastructure


Requirements

  • 4+ years of experience as a Data Engineer in high-throughput environments such as trading, crypto, or fintech

  • Expert-level Python (pandas, pyarrow, asyncio) and SQL skills, with strong fundamentals in algorithms and distributed systems

  • Proven experience with streaming frameworks (Flink, Spark Structured Streaming, Kafka Streams) and message buses (Kafka, Kinesis, Pulsar)

  • In-depth understanding of blockchain data structures (blocks, receipts, logs), indexers (The Graph, Substreams), and node/RPC infrastructure

  • Familiarity with CEX market APIs (REST & WebSocket) and mechanics of perpetual futures (funding, mark price, open interest, liquidations)

  • Proficient in cloud-native development (AWS or GCP), including IaC (Terraform/CDK), CI/CD, and container orchestration (EKS/GKE)

  • A strong track record of building and owning production systems end-to-end, with clear documentation and operational rigor

  • Passionate about perpetual futures and market microstructure—you don’t need to be a trader, but curiosity is key


What We Offer

  • Flexible Working Hours – Enjoy autonomy over your schedule

  • Generous Vacation Policy – 25 days vacation per year plus public holidays

  • Competitive Salary – With regular performance reviews

  • Token Allocation – Be rewarded with tokens as part of our compensation package

  • Growth Opportunities – Be part of an exciting new project with significant career growth potential

  • Innovative Work Culture – Join a team that’s at the cutting edge of Web3, AI, and DeFi, and help shape the future of the digital economy

  • Fun and Engaging Team Activities – Game nights, virtual celebrations, and work retreats to keep things exciting

Apply now Apply later

* Salary range is an estimate based on our AI, ML, Data Science Salary Index 💰

Job stats:  1  0  0
Category: Engineering Jobs

Tags: APIs AWS BigQuery Blockchain CI/CD Crypto Data quality dbt Distributed Systems ETL FinTech Firehose Flink GCP GraphQL Kafka Kinesis NLP Pandas Pipelines Pulsar Python Snowflake Spark SQL Streaming Terraform UX

Perks/benefits: Career development Competitive pay Flex hours Flex vacation Team events

Region: Remote/Anywhere

More jobs like this