Senior Datalake Implementation Specialist

4 - 9 years

5 - 9 Lacs

Posted:10 hours ago| Platform: Naukri logo

Apply

Work Mode

Work from Office

Job Type

Full Time

Job Description


Job TitleSenior DataLake Implementation Specialist
Experience:1012+ Years
Location:Bangalore
Type:Full-time / Contract
Notice Period:Immediate
Job Summary:
We are looking for a highly experienced and sharpDataLake Implementation Specialistto lead and execute scalable data lake projects using technologies such as
Apache Hudi, Hive, Python, Spark, Flink, and
cloud-native tools on AWS or Azure. The ideal candidate must have deep expertise in designing and optimizing modern data lake architectures with strong programming skills and data engineering capabilities.
Key Responsibilities:
  • Design, develop, and implement robust
    data lake architectureson cloud platforms (AWS/Azure).
  • Implement
    streaming and batch data pipelinesusing
    Apache Hudi, Apache Hive, and cloud-native services like
    AWS Glue,
    Azure Data Lake, etc.
  • Architect and optimize ingestion, compaction, partitioning, and indexing strategies in
    Apache Hudi.
  • Develop scalable data transformation and ETL frameworks using
    Python,
    Spark, and
    Flink.
  • Work closely with DataOps/DevOps to build CI/CD pipelines and monitoring tools for data lake platforms.
  • Ensure data governance, schema evolution handling, lineage tracking, and compliance.
  • Collaborate with analytics and BI teams to deliver clean, reliable, and timely datasets.
  • Troubleshoot performance bottlenecks in big data processing workloads and pipelines.

  • Must-Have


    Skills:

  • 4+ yearshands-on experience in
    Data Lake and Data Warehousingsolutions

  • 3+ yearsexperience with
    Apache Hudi, including insert/upsert/delete workflows, clustering, and compaction strategies
  • Strong hands-on experience in
    AWS Glue,
    AWS Lake Formation, or
    Azure Data Lake / Synapse

  • 6+ yearsof coding experience in
    Python, especially in data processing

  • 2+ yearsworking experience in
    Apache Flinkand/or
    Apache Spark
  • Sound knowledge of
    Hive,
    Parquet/ORC formats, and
    DeltaLake vs Hudi vs Iceberg
  • Strong understanding of
    schema evolution,
    data versioning, and
    ACID guaranteesin data lakes

  • Nice to Have:
  • Experience with
    Apache Iceberg,
    Delta Lake
  • Familiarity with
    Kinesis,
    Kafka, or any streaming platform
  • Exposure to
    dbt,
    Airflow, or
    Dagster
  • Experience in
    data cataloging,
    data governance tools, and
    column-level lineage tracking

  • Education & Certifications:
  • Bachelors or Masters degree in Computer Science, Information Technology, or related field
  • Relevant certifications in
    AWS Big Data,
    Azure Data Engineering, or
    Databricks

  • Mock Interview

    Practice Video Interview with JobPe AI

    Start Python Interview
    cta

    Start Your Job Search Today

    Browse through a variety of job opportunities tailored to your skills and preferences. Filter by location, experience, salary, and more to find your perfect fit.

    Job Application AI Bot

    Job Application AI Bot

    Apply to 20+ Portals in one click

    Download Now

    Download the Mobile App

    Instantly access job listings, apply easily, and track applications.

    coding practice

    Enhance Your Python Skills

    Practice Python coding challenges to boost your skills

    Start Practicing Python Now
    Enterprise Minds logo
    Enterprise Minds

    Consulting

    Tech City

    RecommendedJobs for You