PySpark Developer

0 years

0.0 Lacs P.A.

Hyderabad, Telangana, India

Posted:5 days ago| Platform: Linkedin logo

Apply Now

Skills Required

pysparkdeveloperdataprocessingapachesparkpythonengineeringawsetlscriptingautomationschedulingairflowunixlinuxhadoophivecapturetuningtroubleshootingwritingtestintegrationpycharmdevelopmentdebuggingagilejenkinsgitprovisioningreportinginformaticatableauqlikviewaimlmodelapisql

Work Mode

On-site

Job Type

Contractual

Job Description

Role : PySpark Developer Locations : Hyderabad & Bangalore Work Mode : Hybrid Interview Mode : Virtual (2 Rounds) Type : Contract-to-Hire (C2H) Job Summary We are looking for a skilled PySpark Developer with hands-on experience in building scalable data pipelines and processing large datasets. The ideal candidate will have deep expertise in Apache Spark , Python , and working with modern data engineering tools in cloud environments such as AWS . Key Skills & Responsibilities Strong expertise in PySpark and Apache Spark for batch and real-time data processing. Experience in designing and implementing ETL pipelines, including data ingestion, transformation, and validation. Proficiency in Python for scripting, automation, and building reusable components. Hands-on experience with scheduling tools like Airflow or Control-M to orchestrate workflows. Familiarity with AWS ecosystem, especially S3 and related file system operations. Strong understanding of Unix/Linux environments and Shell scripting. Experience with Hadoop, Hive, and platforms like Cloudera or Hortonworks. Ability to handle CDC (Change Data Capture) operations on large datasets. Experience in performance tuning, optimizing Spark jobs, and troubleshooting. Strong knowledge of data modeling, data validation, and writing unit test cases. Exposure to real-time and batch integration with downstream/upstream systems. Working knowledge of Jupyter Notebook, Zeppelin, or PyCharm for development and debugging. Understanding of Agile methodologies, with experience in CI/CD tools (e.g., Jenkins, Git). Preferred Skills Experience in building or integrating APIs for data provisioning. Exposure to ETL or reporting tools such as Informatica, Tableau, Jasper, or QlikView. Familiarity with AI/ML model development using PySpark in cloud environments Skills: ci/cd,zeppelin,pycharm,pyspark,etl tools,control-m,unit test cases,tableau,performance tuning,jenkins,qlikview,informatica,jupyter notebook,api integration,unix/linux,git,aws s3,hive,cloudera,jasper,airflow,cdc,pyspark, apache spark, python, aws s3, airflow/control-m, sql, unix/linux, hive, hadoop, data modeling, and performance tuning,agile methodologies,aws,s3,data modeling,data validation,ai/ml model development,batch integration,apache spark,python,etl pipelines,shell scripting,hortonworks,real-time integration,hadoop Show more Show less

Technology Solutions
Tech City

RecommendedJobs for You

Hyderabad, Telangana, India

Hyderabad, Telangana, India

Hyderabad, Telangana, India

Hyderabad, Telangana, India