Intermediate Data Engineer - OP01505-00

Astana, Astana, Kazakhstan - Remote

Applications have closed

Dev.Pro

Globally distributed 850+ tech talent-rich software development partner. Result driven. Quality obsessed. Scale your business with Dev.Pro

View all jobs at Dev.Pro

🟒 Join Dev.Pro - a US-based outsourcing software development company that has been delivering exceptional software experience to clients since 2011.

We invite a talented, organized, and reliable Data Engineer to join our team and play an important role in building a project that connects the art market with the digital-first financial ecosystem. Work with a global leader in artwork tokenization and blockchain-secured provenance, using your skills to contribute to innovative projects at the crossroads of the blockchain and art industries.

🟩 What's in it for you:

  • Engage in an exciting blockchain-related project transforming the art market
  • Work within a unique business domain that collaborates with world-class brands
  • Join a dynamic no-legacy startup and contribute to building processes, tools, and products from scratch

βœ… Is that you?

  • 4+ years of experience in data engineering, encompassing data extraction, transformation, and migration.
  • Advanced experience with data extraction from unstructured files and legacy systems.
  • Proven expertise in migrating data from file-based storage systems to cloud storage solutions, ideally on Google Cloud Platform.
  • Proficiency with relational databases, specifically MariaDB or MySQL, as well as cloud-native solutions such as Google Cloud Storage, Google BigQuery, and optionally Snowflake or Amazon Redshift.
  • Strong programming skills in Python, with a focus on data manipulation, automation, and re-implementing custom tools.
  • Extensive experience with ETL/ELT pipeline development and workflow orchestration tools (e.g., Apache Airflow, Luigi, Google Dataflow, Prefect).
  • Hands-on experience with batch processing frameworks and real-time data processing frameworks.
  • Experience with data pipeline development using programming languages, including batch processing implementation.
  • In-depth understanding of data modeling, data warehousing, and best practices for designing scalable data architectures.
  • Practical experience in developing or re-engineering data mastering tools for the purpose of data cleaning, standardization, and preparation.
  • Expertise in RDBMS functionalities, such as stored procedures, triggers, partitioning, indexes, and structural changes.
  • Ability to handle Personally Identifiable Information (PII) data within pipelines and data storage systems.
  • Experience with NoSQL databases, such as MongoDB, Cassandra, or HBase.
  • Experience with monitoring tools such as Prometheus, Grafana, and CloudWatch to oversee data pipelines and systems.
  • Knowledge of best practices in database management, performance optimization, data security, and ensuring consistency across distributed systems.
  • Ability to critically evaluate data architecture and provide strategic recommendations for infrastructure improvements.
  • Upper-Intermediate+ English level

Desirable:

  • Familiarity with JavaScript for maintaining or enhancing legacy systems and cross-functional integration.
  • Experience with ElasticSearch for indexing and querying large datasets.
  • Proficiency with analytical tools such as Tableau, Power BI, Looker, or similar platforms for data visualization and insights generation.
  • Interest or background in the art industry, particularly related to digital asset management and tokenization.
  • Demonstrated ability to collaborate in cross-functional teams and contribute to multidisciplinary projects.
  • Experience with PostgreSQL and understanding its application in data engineering environments.
  • Knowledge of specific services related to data engineering, including key metrics and business processes relevant to the industry domain.
  • Experience with MLOps tools and practices to streamline machine learning deployment and operations.
  • Basic understanding of existing machine learning models and algorithms.

🧩Key responsibilities and your contribution

As a Data Engineer, you will contribute significantly to the data team's efforts on a rapidly expanding project, helping to manage increasing workloads while maintaining high standards of quality and efficiency. Your key responsibilities will include:

  • Taking full responsibility for the data warehouse and pipeline, including planning, coding, reviews, and delivery to the production environment.
  • Migrating data from existing file storage systems to the Google Cloud Platform, including Google Cloud Storage and BigQuery.
  • Designing, developing, and maintaining ETL/ELT pipelines to support data migration and integration.
  • Collaborating with team members to re-implement existing custom data mastering tools, with a focus on improving data cleaning and standardization capabilities.
  • Conducting thorough evaluations of the existing technology stack and providing data-driven recommendations for improvements, including re-evaluating database solutions and orchestration tools.
  • Developing a new scraper system to extract and aggregate data from diverse external sources, ensuring integration with existing platforms.
  • Ensuring the integrity, consistency, and quality of data through optimized processes and validation protocols.
  • Working closely with a small, dynamic team to ensure that project milestones are met effectively, with an emphasis on scalability, reliability, and sustainability of solutions.

🎾 What's working at Dev.Pro like?

Working on projects that touch upon the lives of millions globally, Dev.Pro recognizes the importance of people behind the technology and has the most genuine appreciation for their diversity, uniqueness, and individuality. We are building an inclusive, accepting, and friendly work environment where everyone feels supported and cared for. We offer:

βœ”οΈ 30 paid rest days per year to use as holidays/vacation/other on the desired and requested dates
βœ”οΈ Professional and personal growth: ability to shift your career path, access to English lessons and DP University educational flows
βœ”οΈ Healthcare kit to cover visits to the gym, sports nutrition, and similar needs; healthcare insurance fully compensated by Dev.Pro
βœ”οΈ Pregnancy support, maternity and military leave
βœ”οΈ 5 sick leave days, up to 60 days of medical leave, and up to 6 days of leave per year due to family reasons (i.e., wedding/funeral/baby birth)
βœ”οΈ Fun online activities and team-building events

Our next steps:

βœ… Submit a CV in English β€” βœ… Intro call with a Recruiter β€” βœ… Internal interview β€” βœ… Client interview β€” βœ… Offer

Interested? Find out more:

πŸ“‹How we work

πŸ’» LinkedIn Page

πŸ“ˆ Our website

πŸ’»IG Page

* Salary range is an estimate based on our AI, ML, Data Science Salary Index πŸ’°

Job stats:  2  0  0
Category: Engineering Jobs

Tags: Airflow Architecture BigQuery Blockchain Cassandra Dataflow Data pipelines Data visualization Data warehouse Data Warehousing Distributed Systems Elasticsearch ELT Engineering ETL GCP Google Cloud Grafana HBase JavaScript Looker Machine Learning MariaDB ML models MLOps MongoDB MySQL NoSQL Pipelines PostgreSQL Power BI Python RDBMS Redshift Security Snowflake Tableau

Perks/benefits: Career development Medical leave Startup environment Team events

Regions: Remote/Anywhere Asia/Pacific
Country: Kazakhstan

More jobs like this