9 - 12 years
0 - 3 Lacs
Posted:6 days ago|
Platform:
Hybrid
Full Time
Experience - 9 years - 12 years Location - Mumbai / Chennai / Bangalore / Pune Develop and maintain scalable data pipelines using PySpark and Spark SQL for processing large datasets efficiently. Write clean, reusable, and optimized code in Python for data manipulation, analysis, and automation tasks. Design and implement ETL workflows to extract, transform, and load data from various structured and unstructured sources. Collaborate with data engineers, analysts, and stakeholders to understand data requirements and deliver solutions. Optimize Spark jobs for performance tuning, resource utilization, and minimizing execution time. Work with distributed computing frameworks to process and analyze big data in a cloud or on-premises environment. Develop and maintain unit tests to ensure the accuracy and reliability of data pipelines and transformations. Utilize Spark SQL for querying and managing large datasets stored in distributed systems like Hadoop or cloud storage. Monitor and troubleshoot data pipeline issues, ensuring reliability and timely delivery of data. Stay updated with the latest advancements in PySpark, Spark SQL, and big data technologies to improve existing systems.
Hexaware Technologies
Upload Resume
Drag or click to upload
Your data is secure with us, protected by advanced encryption.
My Connections Hexaware Technologies
IT Services and IT Consulting
10001 Employees
513 Jobs
Key People
Pune, Chennai, Bengaluru
0.5 - 3.0 Lacs P.A.
5.0 - 6.0 Lacs P.A.
Hyderabad
10.0 - 20.0 Lacs P.A.
Chennai, Tamil Nadu, India
Experience: Not specified
Salary: Not disclosed
Pune, Chennai, Bengaluru
10.0 - 20.0 Lacs P.A.
Salary: Not disclosed
Gurugram, Bengaluru, Mumbai (All Areas)
5.5 - 15.5 Lacs P.A.
Bengaluru, Delhi / NCR, Mumbai (All Areas)
15.0 - 25.0 Lacs P.A.
Pune, Maharashtra, India
Salary: Not disclosed
Gurgaon, Haryana, India
Salary: Not disclosed