Pyspark Developer

5 - 10 years

5 - 15 Lacs

Posted:16 hours ago| Platform: Naukri logo

Apply

Work Mode

Work from Office

Job Type

Full Time

Job Description

Location:

Experience:

Key Responsibilities:

  • Design and build robust, scalable ETL/ELT pipelines using

    PySpark

    to ingest data from diverse sources (databases, logs, APIs, files).
  • Transform and curate raw transactional and log data into analysis-ready datasets in the Data Hub and analytical data marts.
  • Develop reusable and parameterized Spark jobs for batch and micro-batch processing.
  • Optimize performance and scalability of PySpark jobs across large data volumes.
  • Ensure data quality, consistency, lineage, and proper documentation across ingestion flows.
  • Collaborate with Data Architects, Modelers, and Data Scientists to implement ingestion logic aligned with business needs.
  • Work with cloud-based data platforms (e.g.,

    AWS S3, Glue, EMR, Redshift

    ) for data movement and storage.
  • Support version control,

    CI/CD

    , and infrastructure-as-code where applicable

Required Skills & Qualifications:

  • 5+ years of experience in data engineering, with strong focus on

    PySpark/Spark

    for big data processing.
  • Expertise in building data pipelines and ingestion frameworks from relational, semi-structured

    (JSON, XML)

    , and unstructured sources

    (logs, PDFs).

  • Proficiency in

    Python

    with strong knowledge of data processing libraries.
  • Strong

    SQL

    skills for querying and validating data in platforms like

    Amazon Redshift, PostgreSQL

    , or similar.
  • Experience with distributed computing frameworks (e.g.,

    Spark on EMR, Databricks

    ).
  • Familiarity with workflow orchestration tools (e.g.,

    AWS Step Functions

    , or similar).
  • Solid understanding of

    data lake / data warehouse architectures

    and

    data modeling basics.

Preferred Qualifications:

  • Experience with AWS data services:

    Glue, S3, Redshift, Lambda, CloudWatch, etc.

  • Familiarity with

    Delta Lake

    or similar for large-scale data storage.
  • Exposure to real-time streaming frameworks (e.g.,

    Spark Structured Streaming, Kafka

    ).
  • Knowledge of data governance, lineage, and cataloging tools (e.g.,

    AWS Glue Catalog, Apache Atlas

    ).
  • Understanding of DevOps/CI-CD pipelines for data projects using Git, Jenkins, or similar tools.

Application Process:

Interested candidates, email resume and cover letter to velkiruba.s@sunware.in

Mock Interview

Practice Video Interview with JobPe AI

Start PySpark Interview
cta

Start Your Job Search Today

Browse through a variety of job opportunities tailored to your skills and preferences. Filter by location, experience, salary, and more to find your perfect fit.

Job Application AI Bot

Job Application AI Bot

Apply to 20+ Portals in one click

Download Now

Download the Mobile App

Instantly access job listings, apply easily, and track applications.

coding practice

Enhance Your Python Skills

Practice Python coding challenges to boost your skills

Start Practicing Python Now
Sunware Technologies logo
Sunware Technologies

IT Services and IT Consulting

San Diego

RecommendedJobs for You