Data Platform Engineer (Coding & ETL Tooling)
Paris, France
⚠️ We'll shut down after Aug 1st - try foo🦍 for all jobs in tech ⚠️
Blackfluo.ai
Data Platform Engineer (Coding & ETL Tooling)
Position Overview
We are seeking a Data Platform Engineer with a strong background in modern coding environments and open-source ETL/ELT technologies. The successful candidate will support the development, orchestration, and automation of data workflows using tools like Python, R, GitLab Runners, Airflow, and dbt. This role also involves managing and optimizing collaborative development environments (GitHub, GitLab) and supporting IDE usage across data science and engineering teams.
Key Responsibilities
Coding Environment Management
- Support the setup and maintenance of development environments using IDEs such as VSCode, RStudio, Cursor, and Jupyter
- Enable best practices for collaborative coding in languages such as Python, R, and Stata
- Ensure integration between IDEs, data platforms, and source control tools for streamlined workflows
- Assist in optimizing development environments for reproducibility, package management, and dependency tracking
Source Control & CI/CD
- Administer Git-based version control systems (GitHub, GitLab), including branching strategies, access control, and repo management
- Develop and manage CI/CD pipelines using GitLab Runners and GitHub Actions for data pipelines and analytical code
- Promote code quality through automated testing, linting, and review workflows
- Support onboarding and upskilling of users in Git workflows and coding standards
ETL/ELT Tooling & Orchestration
- Design and implement data transformation pipelines using open-source tools like Apache Airflow, dbt, and VTL (Validation and Transformation Language)
- Maintain orchestration workflows and monitor execution of scheduled jobs
- Optimize task dependencies, retries, and performance within Airflow DAGs and dbt models
- Integrate ETL tools with source systems, metadata layers, and data warehouses
Automation & Reproducibility
- Build reproducible workflows for data science, statistical analysis, and reporting using templated code bases and configuration-driven pipelines
- Develop modular, reusable components for data ingestion, cleaning, validation, and transformation
- Create infrastructure-as-code templates for deploying ETL tools in cloud or on-prem environments
- Support interoperability and standardization across analytics and data engineering teams
Required Qualifications
Technical Skills
- 6+ years of experience with data scripting and statistical programming languages (Python, R, Stata)
- Strong proficiency with Git-based workflows and tools (GitLab, GitHub, GitHub Actions)
- Experience configuring and working within IDEs such as VSCode, RStudio, Jupyter, and/or Cursor
- Proven track record implementing and managing open-source ETL/ELT tools (Airflow, dbt, GitLab Runners, VTL)
- Familiarity with data orchestration, testing, and observability for pipelines
DevOps & Workflow Automation
- Experience developing CI/CD pipelines for analytical and data engineering use cases
- Knowledge of containerization (Docker) and task execution environments (Kubernetes, GitLab Runners)
- Scripting expertise (Bash, Python, YAML) for configuration, automation, and job orchestration
- Understanding of software engineering best practices (modular design, unit testing, reproducibility)
Preferred Qualifications
- Bachelors or Masters degree in Data Engineering, Computer Science, Statistics, or related field
- Experience working in collaborative research or analytics teams with reproducible coding standards
- Knowledge of data validation frameworks (e.g., Great Expectations, VTL), metadata integration, and lineage tracking
- Familiarity with cloud-native infrastructure and deployment (AWS, GCP, Azure)
- Contributions to or experience working with open-source ETL/analytics tooling
* Salary range is an estimate based on our AI, ML, Data Science Salary Index 💰
Job stats:
1
1
0
Category:
Engineering Jobs
Tags: Airflow AWS Azure CI/CD Computer Science Data pipelines dbt DevOps Docker ELT Engineering ETL GCP Git GitHub GitLab Jupyter Kubernetes Open Source Pipelines Python R Research Stata Statistics Testing
Region:
Europe
Country:
France
More jobs like this
Explore more career opportunities
Find even more open roles below ordered by popularity of job title or skills/products/technologies used.
Sr. Data Engineer jobsPrincipal Data Engineer jobsBusiness Intelligence Developer jobsPower BI Developer jobsData Scientist II jobsStaff Data Scientist jobsPrincipal Software Engineer jobsStaff Machine Learning Engineer jobsDevOps Engineer jobsData Science Intern jobsJunior Data Analyst jobsAI/ML Engineer jobsSoftware Engineer II jobsStaff Software Engineer jobsData Science Manager jobsData Manager jobsLead Data Analyst jobsData Analyst Intern jobsData Specialist jobsSr. Data Scientist jobsBusiness Data Analyst jobsData Governance Analyst jobsBusiness Intelligence Analyst jobsData Engineer III jobsSenior Backend Engineer jobs
Business Intelligence jobsAirflow jobsMLOps jobsOpen Source jobsKafka jobsEconomics jobsKPIs jobsGitHub jobsLinux jobsJavaScript jobsTerraform jobsRAG jobsPostgreSQL jobsBanking jobsPrompt engineering jobsStreaming jobsScikit-learn jobsData Warehousing jobsNoSQL jobsRDBMS jobsClassification jobsComputer Vision jobsPhysics jobsdbt jobsPandas jobs
Google Cloud jobsScala jobsHadoop jobsLangChain jobsGPT jobsData warehouse jobsMicroservices jobsR&D jobsBigQuery jobsCX jobsDistributed Systems jobsELT jobsReact jobsScrum jobsOracle jobsLooker jobsIndustrial jobsPySpark jobsOpenAI jobsJira jobsRobotics jobsRedshift jobsSAS jobsUnstructured data jobsTypeScript jobs