Aws Data Engineer

5 - 8 years

10 - 20 Lacs

Posted:20 hours ago| Platform: Naukri logo

Apply

Work Mode

Work from Office

Job Type

Full Time

Job Description

Job Overview

Primary Responsibilities

  • Design, develop, and maintain data lakes and data pipelines on AWS using ETL frameworks and Databricks.
  • Integrate and transform large-scale data from multiple heterogeneous sources into a centralized data lake environment.
  • Implement and manage Delta Lake architecture using Databricks Delta or Apache Hudi.
  • Develop end-to-end data workflows using PySpark, Databricks Notebooks, and Python scripts for ingestion, transformation, and enrichment.
  • Design and develop data warehouses and data marts for analytical workloads using Snowflake, Redshift, or similar systems.
  • Design and evaluate data models (Star, Snowflake, Flattened) for analytical and transactional systems.
  • Optimize data storage, query performance, and cost across the AWS and Databricks ecosystem.
  • Build and maintain CI/CD pipelines for Databricks notebooks, jobs, and Python-based data processing scripts.
  • Collaborate with data scientists, analysts, and stakeholders to deliver high-performance, reusable data assets.
  • Maintain and manage code repositories (Git) and promote best practices in version control, testing, and deployment.
  • Participate in making major technical and architectural decisions for data engineering initiatives.
  • Monitor and troubleshoot Databricks clusters, Spark jobs, and ETL processes for performance and reliability.
  • Coordinate with business and technical teams through all phases of the software development life cycle.

You Must Have

  • 5+ years of experience building and managing Data Lake Architecture on AWS Cloud
  • 3+ years of experience with AWS Data services such as S3, Glue, Lake Formation, EMR, Kinesis, RDS, DMS, and Redshift.
  • 3+ years of experience building Data Warehouses on Snowflake, Redshift, HANA, Teradata, or Exasol.
  • 3+ years of hands-on experience working with Apache Spark or PySpark, on Databricks.
  • 3+ years of experience implementing Delta Lakes using Databricks Delta or Apache Hudi.
  • 3+ years of experience in ETL development using Databricks, AWS Glue, or other modern frameworks.
  • Proficiency in Python for data engineering, automation, and API integrations.
  • Experience in Databricks Jobs, Workflows, and Cluster Management.
  • Experience with CI/CD pipelines and Infrastructure as Code (IaC) tools like Terraform or CloudFormation is a plus.
  • Bachelors degree in computer science, Information Technology, Data Science, or related field.
  • Experience working on Agile projects and methodology in general.

We Value

  • Strong SQL, RDBMS, and data modeling skills.
  • Experience with Databricks Unity Catalog, Delta Live Tables (DLT), and MLflow for data governance and model lifecycle.
  • AWS or Databricks Cloud Certifications (e.g., AWS Data Analytics Specialty, Databricks Certified Data Engineer Professional) are a big plus.
  • Understanding data security, access control, and compliance in cloud environments.
  • Strong analytical, problem-solving, and communication skills

Mock Interview

Practice Video Interview with JobPe AI

Start Python Interview
cta

Start Your Job Search Today

Browse through a variety of job opportunities tailored to your skills and preferences. Filter by location, experience, salary, and more to find your perfect fit.

Job Application AI Bot

Job Application AI Bot

Apply to 20+ Portals in one click

Download Now

Download the Mobile App

Instantly access job listings, apply easily, and track applications.

coding practice

Enhance Your Python Skills

Practice Python coding challenges to boost your skills

Start Practicing Python Now
Celebal Technologies logo
Celebal Technologies

Technology Consulting and Services

Ahmedabad

RecommendedJobs for You

hyderabad, pune, bengaluru