AWS Databricks - Lead

8 - 12 years

12 - 18 Lacs

Posted:2 weeks ago| Platform: Naukri logo

Apply

Work Mode

Work from Office

Job Type

Full Time

Job Description

General Roles & Responsibilities:

  • Technical Leadership: Demonstrate leadership, and ability to guide business and technology teams in adoption of best practices and standards
  • Design & Development: Design, develop, and maintain robust, scalable, and high-performance data estate
  • Architecture: Architect and design robust data solutions that meet business requirements & include scalability, performance, and security.
  • Quality: Ensure the quality of deliverables through rigorous reviews, and adherence to standards.
  • Agile Methodologies: Actively participate in agile processes, including planning, stand-ups, retrospectives, and backlog refinement.
  • Collaboration: Work closely with system architects, data engineers, data scientists, data analysts, cloud engineers and other business stakeholders to determine optimal solution & architecture that is future-proof too.
  • Innovation: Stay updated with the latest industry trends and technologies, and drive continuous improvement initiatives within the development team.
  • Documentation: Create and maintain technical documentation, including design documents, and architectural user guides.

Technical Responsibilities:

  • Optimize data pipelines for performance and efficiency.
  • Work with Databricks clusters and configuration management tools.
  • Use appropriate tools in the cloud data lake development and deployment. Developing/implementing cloud infrastructure to support current and future business needs.
  • Provide technical expertise and ownership in the diagnosis and resolution of issues.
  • Ensure all cloud solutions exhibit a higher level of cost efficiency, performance, security, scalability, and reliability.
  • Manage cloud data lake development and deployment on AWSDatabricks.
  • Manage and create workspaces, configure cloud resources, view usage data, and manage account identities, settings, and subscriptions in Databricks

Required Technical Skills:

  • Experience & Proficiency with Databricks platform - Delta Lake storage, Spark (PySpark, Spark SQL). Must be well versed with Databricks Lakehouse, Unity Catalog concept and its implementation in enterprise environments.
  • Familiarity of data design pattern - medallion architecture to organize data in a Lakehouse.
  • Experience & Proficiency with AWS Data Services S3, Glue, Athena, Redshift etc.| Airflow scheduling
  • Proficiency in SQL and experience with relational databases.
  • Proficiency in at least one programming language (e.g., Python, Java) for data processing and scripting.
  • Experience with DevOps practices - AWS DevOps for CI/CD, Terraform/CDK for infrastructure as code
  • Good understanding of

    data principles, Cloud Data Lake design & development including data ingestion, data modeling and data distribution.

  • Jira:

    Proficient in using Jira for managing projects and tracking progress.

Other Skills:

  • Strong communication and interpersonal skills.
  • Collaborate with data stewards, data owners, and IT teams for effective implementation
  • Understanding of business processes and terminology preferably

    Logistics

  • Experienced with Scrum and Agile Methodologies

Qualification

  • Bachelors degree in information technology or a related field. Equivalent experience may be considered.
  • Overall experience of

    8-12 years

    in Data Engineering

Mandatory Competencies
Data Science and Machine Learning - Data Science and Machine Learning - Databricks
Data on Cloud - Azure Data Lake (ADL)
Agile - Agile
Data Analysis - Data Analysis
Big Data - Big Data - Pyspark
Data on Cloud - AWS S3
Data on Cloud - Redshift
ETL - ETL - AWS Glue
Python - Python
DevOps - CI/CD
Beh - Communication and collaboration
Cloud - Azure - Azure Data Factory (ADF), Azure Databricks, Azure Data Lake Storage, Event Hubs, HDInsight
Database - Database Programming - SQL
Agile - Agile - SCRUM
QA/QE - QA Analytics - Data Analysis
Cloud - AWS - AWS S3, S3 glacier, AWS EBS
Cloud - AWS - Tensorflow on AWS, AWS Glue, AWS EMR, Amazon Data Pipeline, AWS Redshift
Programming Language - Python - Python Shell
Development Tools and Management - Development Tools and Management - CI/CD
Cloud - AWS - AWS Lambda,AWS EventBridge, AWS Fargate

Mock Interview

Practice Video Interview with JobPe AI

Start PySpark Interview
cta

Start Your Job Search Today

Browse through a variety of job opportunities tailored to your skills and preferences. Filter by location, experience, salary, and more to find your perfect fit.

Job Application AI Bot

Job Application AI Bot

Apply to 20+ Portals in one click

Download Now

Download the Mobile App

Instantly access job listings, apply easily, and track applications.

coding practice

Enhance Your Skills

Practice coding challenges to boost your skills

Start Practicing Now
Iris Software logo
Iris Software

Information Technology & Services

Centennial

RecommendedJobs for You

Pune, Maharashtra, India

Chennai, Tamil Nadu, India