Cloud (AWS) / Databricks Technical Architect
United States-%LABEL POSITION TYPE REMOTE ANY%
Kaizen Analytix
Rapid Delivery, Continuous Improvement | We understand that analytics is much more than just the math. It’s the story behind the data. We help businesses understand what happened and show how they can learn from these outcomes to predict and...Job Summary:
We are seeking an experienced and dynamic Cloud (AWS) / Databricks Technical Architect to join our team. The ideal candidate will have extensive expertise in designing, building, and deploying cloud-based solutions using AWS and Databricks, with a focus on data engineering, analytics, and machine learning. In this role, you will be responsible for driving the architecture and implementation of scalable, secure, and efficient cloud-based data solutions that support the organization's data-driven initiatives.
Key Responsibilities:
- Lead the design and implementation of cloud-based data architectures on AWS, utilizing a wide range of AWS services (e.g., S3, EC2, Lambda, RDS, Redshift, Athena, Glue).
- Architect and deploy scalable, secure, and cost-effective Databricks environments to process large volumes of data for analytics, data engineering, and machine learning.
- Provide leadership in designing modern data architectures, including real-time data pipelines, ETL/ELT workflows, and big data processing systems using Databricks and AWS technologies.
- Define and implement best practices for managing and optimizing data lakes, data warehouses, and data pipelines.
- Ensure architecture decisions align with business requirements, security policies, and compliance standards.
Requirements
- AWS Certification: Maintain and leverage AWS certification to design and implement cloud solutions.
- Cloud Architecture: Design and optimize cloud architecture for scalability and efficiency.
- Containers & Orchestration: Implement containers and orchestration tools for streamlined application deployment.
- Microservices Architecture: Design and manage microservices architecture for flexible and scalable systems.
- Cloud Environment Setup and Configurations: Set up and configure cloud environments to meet project requirements.
- Security & Access Management: Ensure secure access management and compliance within cloud environments.
- SQL, Python, Visualization & Analytical Tools: Use SQL, Python, and analytical tools for data processing and visualization.
- API Development & Management: Develop and manage APIs for seamless data integration and functionality.
Education & Experience:
- Bachelor’s or Master’s degree in Computer Science, Information Technology, Data Engineering, or a related field.
- 8+ years of experience in cloud architecture, data engineering, or a similar technical role, with at least 5 years of hands-on experience working with AWS and Databricks.
- Proven track record in architecting and deploying large-scale data engineering solutions on AWS and Databricks.
- Experience working with various data processing frameworks (e.g., Apache Spark, Apache Kafka, Airflow) and cloud-based data storage solutions.
Technical Skills & Competencies:
- Deep expertise in AWS services, including but not limited to S3, EC2, Lambda, Glue, Redshift, Athena, and RDS.
- Strong experience with Databricks, including notebook creation, Spark-based processing, and managing Databricks clusters.
- Expertise in data engineering concepts, including ETL/ELT, data lakes, data pipelines, and real-time streaming architectures.
- Proficiency in programming languages such as Python, Scala, SQL, or Java for data processing and solution development.
- Experience with DevOps practices, CI/CD pipelines, containerization (e.g., Docker, Kubernetes), and infrastructure as code (e.g., Terraform, CloudFormation).
- Familiarity with machine learning workflows and tools, particularly those that integrate with Databricks (e.g., MLflow, Spark MLlib).
- Strong understanding of cloud security best practices, including IAM, encryption, and network security.
Preferred Qualifications:
- Experience with big data frameworks (e.g., Hadoop, Spark) and container orchestration platforms (e.g., Kubernetes).
- Familiarity with data governance, privacy, and compliance frameworks.
* Salary range is an estimate based on our AI, ML, Data Science Salary Index 💰
Tags: Airflow API Development APIs Architecture Athena AWS Big Data CI/CD CloudFormation Computer Science Databricks Data governance Data pipelines DevOps Docker EC2 ELT Engineering ETL Hadoop Java Kafka Kubernetes Lambda Machine Learning Microservices MLFlow Pipelines Privacy Python Redshift Scala Security Spark SQL Streaming Terraform
Perks/benefits: Flex hours
More jobs like this
Explore more career opportunities
Find even more open roles below ordered by popularity of job title or skills/products/technologies used.