Senior Data Engineer(SSE)

Hyderabad, India

Ninja Van

Ninja Van is Southeast Asia’s leading logistics provider, with the highest service coverage over 6 countries in the region. Experience the joy of hassle-free deliveries by shipping with Ninja Van today.

View all jobs at Ninja Van

Apply now Apply later

Ninja Van is a late-stage logtech startup that is disrupting a massive industry with innovation and cutting edge technology. Launched 2014 in Singapore, we have grown rapidly to become one of Southeast Asia's largest and fastest-growing express logistics companies. Since our inception, we’ve delivered to 100 million different customers across the region with added predictability, flexibility and convenience. Join us in our mission to connect shippers and shoppers across Southeast Asia to a world of new possibilities. 
More about us: - We process 250 million API requests and 3TB of data every day.- We deliver more than 2 million parcels every day.- 100% network coverage with 2600+ hubs and stations in 6 SEA markets (Singapore, Malaysia, Indonesia, Thailand, Vietnam and Philippines), reaching 500 million consumers.- 2 Million active shippers in all e-commerce segments, from the largest marketplaces to the individual social commerce sellers.- Raised more than US$500 million over five rounds.
We are looking for world-class talent to join our crack team of engineers, product managers and designers. We want people who are passionate about creating software that makes a difference to the world. We like people who are brimming with ideas and who take initiative rather than wait to be told what to do. We prize team-first mentality, personal responsibility and tenacity to solve hard problems and meet deadlines. As part of a small and lean team, you will have a very direct impact on the success of the company.

Roles & Responsibilities

  • Design, develop, and maintain Ninja Van's infrastructure for data streaming, processing, and storage.
  • Build tools to ensure effective maintenance and monitoring of the data infrastructure.
  • Contribute to key architectural decisions for data pipelines and lead the implementation of major initiatives.
  • Collaborate with stakeholders to deliver scalable and high-performance solutions for data requirements, including extraction, transformation, and loading (ETL) from diverse data sources.
  • Enhance the team's data capabilities by sharing knowledge, enforcing best practices, and promoting data-driven decision-making.
  • Develop and enforce Ninja Van's data retention policies and backup strategies, ensuring data is stored redundantly and securely.

Requirements

  • At least 5+ years of experience in a similar role, with a proven track record of building scalable and high-performance data infrastructure using Python, PySpark, Spark, and Airflow.
  • Solid computer science fundamentals, excellent problem-solving skills, and a strong understanding of distributed computing principles.
  • Expert-level SQL knowledge and extensive experience working with both relational and NoSQL databases.
  • Advanced knowledge of Apache Kafka, along with demonstrated proficiency in Hadoop v2, HDFS, and MapReduce.
  • Hands-on experience with stream-processing systems (e.g., Storm, Spark Streaming), big data querying tools (e.g., Pig, Hive, Spark), and data serialization frameworks (e.g., Protobuf, Thrift, Avro).
  • [Good to have] Familiarity with infrastructure-as-code technologies like Terraform, Terragrunt, Ansible, or Helm. Don’t worry if you don’t have this experience—what matters is your interest in learning!
  • [Good to have] Experience with Change Data Capture (CDC) technologies such as Maxwell or Debezium.
  • Bachelor’s or Master’s degree in Computer Science or a related field from a top university.
Tech StackBackend: Play (Java 8+), Golang, Node.js, Python, FastAPIFrontend: AngularJS, ReactJSMobile: Android, Flutter, React NativeCache: Hazelcast, RedisData storage: MySQL, TiDB, Elasticsearch, Delta LakeInfrastructure monitoring: Prometheus, GrafanaOrchestrator: KubernetesContainerization: Docker, ContainerdCloud Provider: GCP, AWSData pipelines: Apache Kafka, Spark Streaming, Maxwell/Debezium, PySpark, TiCDCWorkflow manager: Apache AirflowQuery engines: Apache Spark, Trino
Submit a job applicationBy applying to the job, you acknowledge that you have read, understood and agreed to our Privacy Policy Notice (the “Notice”) and consent to the collection, use and/or disclosure of your personal data by Ninja Logistics Pte Ltd (the “Company”) for the purposes set out in the Notice. In the event that your job application or personal data was received from any third party pursuant to the purposes set out in the Notice, you warrant that such third party has been duly authorised by you to disclose your personal data to us for the purposes set out in the the Notice.
Apply now Apply later

* Salary range is an estimate based on our AI, ML, Data Science Salary Index 💰

Job stats:  0  0  0
Category: Engineering Jobs

Tags: Airflow Angular Ansible APIs Avro Big Data Computer Science Data pipelines Docker E-commerce Elasticsearch ETL GCP Golang Hadoop HDFS Helm Java Kafka MySQL Node.js NoSQL Pipelines Privacy PySpark Python React Spark SQL Streaming Terraform

Perks/benefits: Startup environment

Region: Asia/Pacific
Country: India

More jobs like this