Automation Data Engineer
Jakarta Selatan, DKI Jakarta, Indonesia
We are currently seeking a passionate Data Engineer to join our intelligent Automation Team and help us finding out the best possible scenario in operating the vast amounts of data and help us make smarter decisions to deliver even better products.
Your primary focus will be to design, implement and forecast the cost of data pipelines and infrastructure for the team to perform ETL, analysis, reporting, visualization and dashboarding of large amount of data. You will also help the team to produce the best quality data for process mining to improve and support Robotic Process Automation. With your help, we want to make the process discovery, monitoring, optimization and modelling become seamless yet reliable. This role will rely heavily on technical skills, creative solutions, thorough documentation, and timely delivery.
Responsibilities:
- Support Data Scientist in deploying model into production.
- Lead a team of Data Engineer to work with the team.
- Design and Manage data architecture and instance migrations to support team’s operations.
- Design and Create ETL pipeline from multiple event logs to be ready to be used for Process Mining Analysis.
- Support Data Scientist, Data analyst and other team to prepare data for visualization, modelling, and suggest the most efficient way to perform the tasks.
- Collaborate with Data Scientist, Data Analyst, RPA Engineer, product management and engineering departments to understand the company needs and devise possible solutions.
- Enhancing data collection procedures to include information that is relevant for building our analytical systems.
- Managing, processing, cleansing, and verifying the integrity of data used for analysis, storage and process.
- Doing ad-hoc task to pull, modify, and store the data to platforms that used by the team.
Requirements:
- Minimum 2 year of experience as a data engineer.
- Strong understanding in Python is a must (other programming language is a plus).
- Have advanced skill in docker for machine learning deployment (API).
- Have basic skill in kubernetes, airflow, docker swarm, or other.
- Experienced in cloud computing, knowing how to design architecture in the cloud is a must.
- Strong experience in code versioning: GIT, bitbucket (command).
- Strong knowledge of data storage and processing platform: ElasticSearch/OpenSearch, SQL platforms, Spark, Kafka.
- Know how to write shell script: bash.
- Know how to write SQL.
- Proficient in English.
- Have knowledge of a MLOps (MLflow) is a plus
* Salary range is an estimate based on our AI, ML, Data Science Salary Index 💰
Tags: Airflow APIs Architecture Bitbucket Data pipelines Docker Elasticsearch Engineering ETL Git Kafka Kubernetes Machine Learning MLFlow MLOps OpenSearch Pipelines Python Robotics RPA Spark SQL
More jobs like this
Explore more career opportunities
Find even more open roles below ordered by popularity of job title or skills/products/technologies used.