Posted:1 day ago|
Platform:
On-site
Contractual
Our Client is a global IT services company headquartered in Southborough, Massachusetts, USA. Founded in 1996, with a revenue of $1.8B, with 35,000+ associates worldwide, specializes in digital engineering, and IT services company helping clients modernize their technology infrastructure, adopt cloud and AI solutions, and accelerate innovation. It partners with major firms in banking, healthcare, telecom, and media.
Our Client is known for combining deep industry expertise with agile development practices, enabling scalable and cost-effective digital transformation. The company operates in over 50 locations across more than 25 countries, has delivery centers in Asia, Europe, and North America and is backed by Baring Private Equity Asia.
Key Responsibilities:
Design, develop, and maintain scalable data pipelines using Apache Spark on Databricks. Write efficient and production-ready PySpark or Scala code for data transformation and ETL processes. Integrate data from various structured and unstructured sources into a unified platform. Implement Delta Lake and manage data versioning, updates, and schema evolution. Optimize data processing workflows for performance, scalability, and cost efficiency. Collaborate with data scientists, analysts, and business stakeholders to deliver high-quality datasets. Implement data quality checks, validation routines, and logging mechanisms. Monitor and debug production jobs using Databricks jobs, notebooks, and clusters. Ensure security, privacy, and compliance standards are met throughout the data lifecycle. Provide guidance and mentorship to junior team members.
Required Skills & Qualifications:
5 to 7 years of experience in Big Data development. Hands-on experience with Databricks (including Workflows, Notebooks, Delta Live Tables, Unity Catalog). Strong programming skills in PySpark and/or Scala. Solid understanding of Delta Lake architecture. Proficient in SQL for data analysis and transformation. Experience with cloud platforms such as Azure (Azure Data Lake, Data Factory, Synapse) or AWS (S3, Glue, Redshift). Familiarity with CI/CD for Databricks deployments (e.g., using GitHub Actions, Azure DevOps). Knowledge of data governance, cataloguing, and security best practices. Experience working in an Agile/Scrum environment.
Preferred Skills:
Experience with Databricks Unity Catalog and Delta Live Tables. Exposure to machine learning workflows in Databricks. Experience with Apache Airflow, Kafka, or other orchestration/messaging tools. Certifications such as Databricks Certified Data Engineer Associate/Professional, Azure, or AWS certification.
People Prime Worldwide
Upload Resume
Drag or click to upload
Your data is secure with us, protected by advanced encryption.
Browse through a variety of job opportunities tailored to your skills and preferences. Filter by location, experience, salary, and more to find your perfect fit.
We have sent an OTP to your contact. Please enter it below to verify.
mahabubnagar
16.0 - 22.5 Lacs P.A.
16.0 - 22.5 Lacs P.A.
16.0 - 22.5 Lacs P.A.
16.0 - 22.5 Lacs P.A.
16.0 - 22.5 Lacs P.A.
16.0 - 22.5 Lacs P.A.
16.0 - 22.5 Lacs P.A.
16.0 - 22.5 Lacs P.A.
16.0 - 22.5 Lacs P.A.
pimpri-chinchwad
16.0 - 22.5 Lacs P.A.