DataZymes - Databricks Platform Engineer - ETL

5 - 8 years

10 - 15 Lacs

Posted:1 day ago| Platform: Naukri logo

Apply

Work Mode

Work from Office

Job Type

Full Time

Job Description

Role Overview :

Seasoned Databricks Senior Platform Engineer who can implement scalable, end-to-end data solutions on Databricks over AWS. This role is pivotal in building the foundational infrastructure for data products, ensuring data quality, and enabling machine learning capabilities

Key Responsibilities :

- Design and implement end-to-end ETL pipelines in Databricks, sourcing data from Amazon S3 and Amazon RDS and other sources.

- Design and deploy Databricks workspaces and clusters optimized for performance, scalability, and cost-efficiency.

- Set up data quality frameworks and anomaly detection mechanisms to ensure reliability and trust in data products.

- Enable machine learning capabilities within Databricks by configuring necessary services, libraries, and integrations.

- Collaborate with data scientists, engineers, and DevOps teams to ensure seamless platform operations.

- Automate deployment and monitoring using CI/CD pipelines, GitHub

- Document architecture, workflows, and operational procedures for platform governance.

Requirements :

Core Platform & Cloud :

- 5+ years of experience in platform engineering roles

- Strong expertise in Databricks on AWS

- Proficient with Amazon S3, Amazon RDS, and IAM roles/policies

- Experience with Delta Lake, Structured Streaming, and Databricks Workflows

ETL & Data Quality :

- Hands-on experience building ETL pipelines using Spark

- Familiarity with data validation, profiling, and anomaly detection frameworks

- Knowledge of tools like Great Expectations or custom validation scripts

ML Platform Enablement :

- Experience setting up ML infrastructure including :

- MLflow for experiment tracking and model registry

- Feature Store for reusable feature pipelines

- Databricks Runtime for ML

- Integration with AWS SageMaker or custom ML libraries

- Understanding of model lifecycle management, though model development is out of scope

DevOps & Automation :

- Proficiency with GitHub, CI/CD pipelines, and Terraform or CloudFormation

Preferred Qualifications :

- Experience working with healthcare data

- Exposure to data mesh or data product architecture principles

Mock Interview

Practice Video Interview with JobPe AI

Start Python Interview
cta

Start Your Job Search Today

Browse through a variety of job opportunities tailored to your skills and preferences. Filter by location, experience, salary, and more to find your perfect fit.

Job Application AI Bot

Job Application AI Bot

Apply to 20+ Portals in one click

Download Now

Download the Mobile App

Instantly access job listings, apply easily, and track applications.

coding practice

Enhance Your Python Skills

Practice Python coding challenges to boost your skills

Start Practicing Python Now
Datazymes Analytics

Data Analytics

San Francisco

RecommendedJobs for You

hyderabad, telangana