Dataiku Developer
INDIA - BENGALURU - BIRLASOFT OFFICE, IN
Birlasoft
At Birlasoft we combine the power of domain, enterprise, and digital technologies to reimagine business potential. Surpassing expectations, breaking convention!Area(s) of responsibility
Job Title: Dataiku Developer
Location:
Pune/Mumbai/Chennai/Hyd/Bangalore/Noida
About the Role:
We are seeking a skilled Dataiku Developer to join our data science workbench program team. The ideal candidate will design, develop, and deploy scalable data pipelines and advanced analytics and AI/ML solutions using the Dataiku platform. You will work closely with data scientists, engineers, and business stakeholders to enable data-driven decision-making.
Key Responsibilities:
- Design, develop, extend and maintain end-to-end data workflows and pipelines in Dataiku DSS.
- Collaborate with data scientists and analysts to operationalize machine learning models.
- Leverage Generative AI models and tools within Dataiku to build advanced AI-powered applications and analytics solutions.
- Integrate Dataiku with various data sources (databases, cloud storage, APIs).
- Develop and optimize SQL queries and Python/R scripts for data extraction and transformation across relational and NoSQL databases
- Work extensively with cloud data warehouses like Amazon Redshift and/or Snowflake for data ingestion, transformation, and analytics.
- Implement automation and scheduling of data workflows.
- Monitor and troubleshoot data pipelines to ensure data quality and reliability.
- Document technical solutions and best practices for data processing and analytics.
Required Skills and Qualifications:
- Proven experience of 4+ years working with Dataiku Data Science Studio (DSS) in a professional environment.
- Strong knowledge of data engineering concepts, ETL/ELT processes.
- Proficiency in Python and/or R for data manipulation and automation.
- Solid SQL skills and experience with relational databases (e.g., MySQL, PostgreSQL, Oracle)
- Familiarity with cloud platforms (AWS or optionally Azure/GCP) and data storage technologies.
- Understanding of machine learning lifecycle and model deployment.
- Experience with REST APIs and integration tools.
- Strong analytical, problem-solving, and communication skills.
- Bachelor’s degree in Computer Science, Engineering, Data Science, or related field.
Preferred Qualifications:
- Experience with big data technologies like Hadoop, Spark, or Kafka and databases like Redshift, Snowflake or AWS S3.
- Knowledge of containerization and orchestration (Docker, Kubernetes).
- Experience with CI/CD pipelines for data projects.
- Prior experience in agile development methodologies.
* Salary range is an estimate based on our AI, ML, Data Science Salary Index 💰
Tags: Agile APIs AWS Azure Big Data CI/CD Computer Science Data pipelines Data quality Docker ELT Engineering ETL GCP Generative AI Hadoop Kafka Kubernetes Machine Learning ML models Model deployment MySQL NoSQL Oracle Pipelines PostgreSQL Python R RDBMS Redshift Snowflake Spark SQL
Perks/benefits: Career development
More jobs like this
Explore more career opportunities
Find even more open roles below ordered by popularity of job title or skills/products/technologies used.