Senior Data Engineer

Latin America#LATAM, Remote, Colombia (Hybrid)

Parser

Elevating organisations through groundbreaking AI innovation. Our services transcend conventional offerings, functioning as dynamic service hubs where innovation, collaboration, and entrepreneurship thrive.

View all jobs at Parser

Apply now Apply later

Senior Data Engineer - LATAM


This position offers you the opportunity to join a fast-growing technology organization that is redefining productivity paradigms in the software engineering industry. Thanks to our flexible, distributed model of global operation and the high caliber of our experts, we have enjoyed triple digit growth over the past five years, creating amazing career opportunities for our people. If you want to accelerate your career working with like-minded subject matter experts, solving interesting problems and building the products of tomorrow, this opportunity is for you.


You'd join the data platform team to develop data infrastructure to support our next generation products and data initiatives.
In this group, you will be responsible for building, operationalizing and optimizing our data and data pipeline architecture, as well as optimizing data flow and collection for cross functional teams.


Do you believe testing should be at the heart of any engineering effort? Do you believe that ruthless simplification and refactoring are the soul of any engineering effort? Do you want to help establish an Engineering culture with these fundamentals? We are looking for individuals hungry to drive change across an already very profitable business. Technology is changing rapidly, and that pace is only accelerating. Advancements in GPT (Generative Pre-trained Transformer) models and Retrieval-Augmented Generation (RAG) algorithms are just a couple examples. We are expanding our Engineering teams to leverage these innovative technologies and modernize our entire Data Management Platform. Come join us to learn and apply modern technologies and prepare our platforms for the many growth opportunities which lie ahead.


 The impact you'll make:


  • Help maintain the existing data pipeline codebase
  • Participate in exploratory and execution any POC
  • Use Git for version control and collaborate with other developers on the team.
  • Expand Unit test coverage.
  • Implement the new data pipeline for the new data source.
  • Adhere to Agile software development methodologies.
  • Work with any data format such as JSON, parquet, avro, XML etc.
  • Use Visual Studio as the primary development environment.
  • Create APIs using cloud services like AWS gateway.


Technology Stack:


Python

AWS/Azure/GCP

Lamdas+Glue

Apache Spark

Scala

Airflow

Databricks

Kafka

Lamdas+Glue



What you'll bring to us:



  • Able to work well within the constructs of an agile development process, including SCRUM, Unit Testing, Continuous Build, and Integration, etc.
  • Proficiency in programming languages, especially Python, is typically required
  • Experience in building and optimizing 'big data' data pipelines, and data sets. This includes designing, constructing, installing, testing, and maintaining highly scalable data management systems
  • Understanding of data structures and algorithms, as well as skills in distributed computing. The ability to develop procedures for data mining, data modeling, and data production.
  • Experience with cloud services such as AWS, Google Cloud, or Azure can be important, as many businesses utilize these platforms for their data storage and processing needs.
  • Familiarity with Databricks delta lake
  • Familiarity with Kafka streaming service.
  • Exceptionally good verbal and written communication skills



Some of the benefits you’ll enjoy working with us:


  • The chance to work in innovative projects with leading brands, that use the latest technologies that fuel transformation.
  • The opportunity to be part of an amazing, multicultural community of tech experts.
  • The opportunity to grow and develop your career with the company. 
  • A flexible and remote working environment. 

 


Come and join our #ParserCommunity. 


Follow us on Linkedin

Apply now Apply later

* Salary range is an estimate based on our AI, ML, Data Science Salary Index 💰

Job stats:  0  0  0
Category: Engineering Jobs

Tags: Agile Airflow APIs Architecture Avro AWS Azure Big Data Databricks Data management Data Mining Data pipelines Engineering GCP Git Google Cloud GPT JSON Kafka Parquet Pipelines Python RAG Scala Scrum Spark Streaming Testing XML

Perks/benefits: Career development Flex hours Startup environment

Regions: Remote/Anywhere South America
Country: Colombia

More jobs like this