Pyspark developer intern

0 years

0 Lacs

Posted:1 day ago| Platform: Linkedin logo

Apply

Work Mode

Remote

Job Type

Internship

Job Description

Key Responsibilities

 
  • Assist in building PySpark ETL pipelines for structured/unstructured healthcare datasets. 
  • Support Delta Lake design ensuring schema evolution and compliance. 
  • Work with mentors on performance tuning and job orchestration. 
  • Document lineage, data quality checks, and compliance processes. 
  • Contribute to CI/CD setup and Git-based workflows. 
  • Participate in testing and validation of data migration outcomes. 
  • Translate business requirements into basic technical artefacts. 
  • Collaborate with senior engineers and project managers on live projects. 

Key Requirements

 

Solid foundation in Python, PySpark, SQL, and cloud data concepts. Exposure to Azure ecosystem (ADF, Databricks, Synapse) is a plus. Passion for learning regulated industry delivery and AI-driven data solutions. 

Preferred Experience


❌ Not suitable for candidates seeking remote/hybrid options. 

❌ Not suitable for candidates without strong fundamentals in Python & SQL. 

Mock Interview

Practice Video Interview with JobPe AI

Start PySpark Interview
cta

Start Your Job Search Today

Browse through a variety of job opportunities tailored to your skills and preferences. Filter by location, experience, salary, and more to find your perfect fit.

Job Application AI Bot

Job Application AI Bot

Apply to 20+ Portals in one click

Download Now

Download the Mobile App

Instantly access job listings, apply easily, and track applications.

coding practice

Enhance Your Python Skills

Practice Python coding challenges to boost your skills

Start Practicing Python Now

RecommendedJobs for You