Senior PySpark Engineer (1 Year Renewable Contract)

8 years

6 - 27 Lacs

Posted:1 month ago| Platform: GlassDoor logo

Apply

Work Mode

On-site

Job Type

Part Time

Job Description

  • Role: Senior PySpark Engineer
  • Experience Required: Minimum 8+ Years
  • Work Location: Hyderabad (5 Days Work from Office)
  • Job Type: Contract to Hire (1 Year/ Renewable)
  • Notice Period: Immediate to 15 Days max
  • Mode of Interview: Virtual

We are seeking a highly skilled PySpark Data Engineer to design, build, and optimize large-scale data pipelines and distributed systems. Beyond deep expertise in Apache Spark (PySpark) and automation, this role requires the ability to manage stakeholders, ensure timely delivery, and assess requirements. You will play a critical role in bridging business needs with technical execution, ensuring high-quality, scalable, and reliable data solutions. Cloudera PySpark experience is preferred.

KEY RESPONSIBILITIES:

  • Architect and guide the refactoring of legacy PySpark scripts into modular, reusable, and configuration-driven frameworks aligned with enterprise standards.
  • Lead migration efforts to Spark 3.3+ and Python 3.10+, ensuring compatibility, performance, and maintainability across distributed systems.
  • Drive modernization by replacing deprecated APIs (e.g., RDDs, legacy UDFs) with efficient DataFrame operations and Pandas UDFs, promoting best practices.
  • Establish and enforce structured logging, robust error handling, and proactive alerting mechanisms for operational resilience.
  • Oversee performance tuning, including partitioning strategies, broadcast joins, and predicate pushdown, to optimize Spark execution plans.
  • Ensure data integrity through schema enforcement, data type consistency, and accurate implementation of Slowly Changing Dimensions (SCD) logic.
  • Collaborate with DevOps and QA teams to integrate Spark workloads into CI/CD pipelines and automated testing frameworks.
  • Mentor and conduct code reviews, providing technical guidance and resolving complex findings to uphold code quality and team growth.
  • Lead performance benchmarking and regression testing initiatives to validate scalability and reliability of Spark applications.
  • Coordinate deployment planning, runbook creation, and production handover, ensuring smooth transitions and operational readiness.
  • Engage with stakeholders to translate business requirements into scalable data processing solutions and contribute to data platform strategy.

Educational Qualification:

  • Graduate/Masters in software engineering/IT/Computer Science or equivalent.

Technical Skills:

PySpark Development (5-7 Years)

  • Refactoring legacy scripts, using DataFrame APIs, avoiding .collect()or equivalent

Spark Optimization (3-5 Years)

  • Broadcast joins, partitioning strategy, predicate pushdown

Pyspark Migration activity (2 Years)

  • Prior experience with Pyspark migration activity.

Testing Frameworks (1+ Years)

  • Pytest, Great Expectations, Deequ for unit/integration/performance testing

Job Type: Contractual / Temporary
Contract length: 12 months

Pay: ₹600,000.00 - ₹2,700,000.00 per year

Work Location: In person

Mock Interview

Practice Video Interview with JobPe AI

Start PySpark Interview
cta

Start Your Job Search Today

Browse through a variety of job opportunities tailored to your skills and preferences. Filter by location, experience, salary, and more to find your perfect fit.

Job Application AI Bot

Job Application AI Bot

Apply to 20+ Portals in one click

Download Now

Download the Mobile App

Instantly access job listings, apply easily, and track applications.

coding practice

Enhance Your Python Skills

Practice Python coding challenges to boost your skills

Start Practicing Python Now

RecommendedJobs for You