Senior Datalake Implementation Specialist

4 - 9 years

5 - 9 Lacs

Posted:None| Platform: Naukri logo

Apply

Work Mode

Work from Office

Job Type

Full Time

Job Description


 Job TitleSenior DataLake Implementation Specialist 
 Experience: 1012+ Years
 Location: Bangalore
 Type: Full-time / Contract
 Notice Period: Immediate
 Job Summary: 
We are looking for a highly experienced and sharp DataLake Implementation Specialist to lead and execute scalable data lake projects using technologies such as Apache Hudi, Hive, Python, Spark, Flink , and cloud-native tools on AWS or Azure . The ideal candidate must have deep expertise in designing and optimizing modern data lake architectures with strong programming skills and data engineering capabilities.
 Key Responsibilities: 
  • Design, develop, and implement robust data lake architectures on cloud platforms (AWS/Azure).
  • Implement streaming and batch data pipelines using Apache Hudi , Apache Hive, and cloud-native services like AWS Glue , Azure Data Lake , etc.
  • Architect and optimize ingestion, compaction, partitioning, and indexing strategies in Apache Hudi .
  • Develop scalable data transformation and ETL frameworks using Python , Spark , and Flink .
  • Work closely with DataOps/DevOps to build CI/CD pipelines and monitoring tools for data lake platforms.
  • Ensure data governance, schema evolution handling, lineage tracking, and compliance.
  • Collaborate with analytics and BI teams to deliver clean, reliable, and timely datasets.
  • Troubleshoot performance bottlenecks in big data processing workloads and pipelines.

  •  Must-Have
    Skills:
  •  
  •  4+ years hands-on experience in Data Lake and Data Warehousing solutions
  •  3+ years experience with Apache Hudi , including insert/upsert/delete workflows, clustering, and compaction strategies
  • Strong hands-on experience in AWS Glue , AWS Lake Formation , or Azure Data Lake / Synapse 
  •  6+ years of coding experience in Python , especially in data processing
  •  2+ years working experience in Apache Flink and/or Apache Spark 
  • Sound knowledge of Hive , Parquet/ORC formats , and DeltaLake vs Hudi vs Iceberg 
  • Strong understanding of schema evolution , data versioning , and ACID guarantees in data lakes

  •  Nice to Have: 
  • Experience with Apache Iceberg , Delta Lake 
  • Familiarity with Kinesis , Kafka , or any streaming platform
  • Exposure to dbt , Airflow , or Dagster 
  • Experience in data cataloging , data governance tools , and column-level lineage tracking 

  •  Education & Certifications: 
  • Bachelors or Masters degree in Computer Science, Information Technology, or related field
  • Relevant certifications in AWS Big Data , Azure Data Engineering , or Databricks 

  • Mock Interview

    Practice Video Interview with JobPe AI

    Start Python Interview
    cta

    Start Your Job Search Today

    Browse through a variety of job opportunities tailored to your skills and preferences. Filter by location, experience, salary, and more to find your perfect fit.

    Job Application AI Bot

    Job Application AI Bot

    Apply to 20+ Portals in one click

    Download Now

    Download the Mobile App

    Instantly access job listings, apply easily, and track applications.

    coding practice

    Enhance Your Python Skills

    Practice Python coding challenges to boost your skills

    Start Practicing Python Now
    Enterprise Minds logo
    Enterprise Minds

    Consulting

    Tech City

    RecommendedJobs for You